Search Results

Search found 9614 results on 385 pages for 'touch mouse'.

Page 146/385 | < Previous Page | 142 143 144 145 146 147 148 149 150 151 152 153  | Next Page >

  • Why doesn't Apache2::SubProcess spawn my subprocess?

    - by codeholic
    The following script works without errors, but /tmp/test.touch is not being created (even being checked later in the command line). It seems to me as if $r->spawn_proc_prog doesn't spawn a process. What may cause the problem? #!/usr/bin/perl use strict; use warnings; use Apache2::RequestUtil; use Apache2::SubProcess (); my $r = Apache2::RequestUtil->request; print "Content-Type: text/plain\n\n"; print eval { $r->spawn_proc_prog('/usr/bin/touch', ['/tmp/test.touch']) } ? `ls -l /tmp/test.touch` : $@;

    Read the article

  • iPhone SDK - Implement Tap to Scroll feature on ImageScrollView

    - by NobodyNobody
    SDK: xCode 3.2 Device: iPhone 3GS OS 3.13 / iPad Case: In Apple Sample Library [ScrollViewSuite]. The [2_Autoscroll] project. In [ThumbImageView], We can drag and drop to scroll the view. So that we can see any menu item by touch the screen and do moving. How to implement this feature in [TagDetectingImageView] ? I have try to copy [touch event] function from [ThumbImageView], modify it and implement to [TagDetectingImageView]. (Just rename the [ThumbImageView] to [TagDetectingImageView] inside those function) But when i touch and move on the [TagDetectingImageView], the view will moved out of the screen (Don't know where is it) Source - Please find the 2_Autoscroll source code in: http://developer.apple.com/iPhone/library/samplecode/ScrollViewSuite/Introduction/Intro.html Thanks

    Read the article

  • iPhone drag/drop

    - by Farid
    Trying to get some basic drag/drop functionality happening for an iPhone application. My current code for trying to do this is as follows: - (void) touchesMoved:(NSSet*)touches withEvent:(UIEvent*)event { UITouch *touch = [touches anyObject]; CGPoint location = [touch locationInView:self]; self.center = location; } This code has the result of a the touched UIView flickering while it follows the touch around. After playing around with it a bit, I also noticed that the UIView seemed to flicker from the 0,0 position on the screen to the currently touched location. Any ideas what I'm doing wrong?

    Read the article

  • send key combination via local network to other mac

    - by choise
    Hi, after a other mac or macbook is connected with a mini display port to a new generation imac you can switch between the imac and the other connected mac pressing cmd+F2(brightness up) using the keyboard of the imac. i connect my macbook to my imac, switch to the macbook and connect the bluetooth keyboard and mouse of my imac to my macbook. after that i'm not able to switch back to my imac, because the keyboard is used on my macbook. i have a second mouse connected to my imac, so i'd would be cool to switch views and control my imac with my second mouse. is it possible to send the shortcut cmd+F2 to my imac from my macbook? via command line or something? using apple script? all ideas are welcome, thanks!

    Read the article

  • Change Ubuntu 11 scrollbars back to the old style

    - by EMP
    The question: I've installed Ubuntu 11.04 and the new scrollbars in Nautilus, GEdit, etc. are driving me insane! How do I go back to the old (Ubuntu 10.10 and earlier) scrollbars? The explanation (for those who haven't upgraded yet): In the Natty Nautilus you can't actually see a scrollbar normally. Instead you mouse over to where it should be (to the left of the border) and then it appears, but it doesn't appear under the mouse pointer! Instead it appears to the side (to the right of the border), so starting to scroll is now a fairly involved mouse manoeuvre. I don't know what UX genius came up with this, but I want none of it.

    Read the article

  • Top/left edges of screen, in Virtualbox, act like bottom/right edges

    - by Ken
    I have Virtualbox running on Windows Vista, and Debian running inside Virtualbox. Everything's running great, for the most part. Everything looks correct. But when I'm in full-screen mode, the top edge seems to act (to the mouse) like it's the bottom edge, and the left edge seems to act like the right edge. For example, if I click in the middle of the desktop and drag left, as if to select some icons, when I hit the very leftmost pixel of the screen, the selection (but not the mouse pointer) jumps to the far right edge of the screen). For the left edge, it's not such a big deal, but not having the top edge is kind of annoying: it means I can't select things from the menu in my top panel by slamming the mouse against the top of the screen. Anyone seen this before? Is there some way to make this work? Thanks!

    Read the article

  • Multi screens are not aligning properly

    - by Steve
    I am using (4) Samsung (SMB2230) screens. When I place my mouse on the first screen at the bottom right and pull it into the second screen the mouse jumps to the middle on the left of the second screen. Then it does the same on the 3rd and 4th screens. The biggest jump is between the 1st and 2nd then it tightens up between 2 and 3 then close between 3 and 4. This is a really pain because my mouse gets hung up and I have to move up and down to get it to change screens.I had (4) Dells screens before and this didn't happen. I'm using Windows 7

    Read the article

  • Context Menu (Right Click) keyboard shortcut in Mac OS X

    - by czerwin
    Is it really possible to invoke a context menu using a keyboard shortcut instead of clicking the right/alt mouse button in OS X? In particular, I would like a menu-key-like feature in OS X. I am wondering whether there is an additional third party software that provides such feature. Please not that the Mouse Keys feature is not an option as I don't want to depend on the position of the mouse cursor. Similar Topics Keyboard Shortcut to Right Click in Mac OS X Right click using keyboard in Mac OS X Enable Right-Click on Mac OS X 10.7.5 by default Keyboard shortcut for spelling dropdown menu in OS X beyond Devonthink Pro? Add application to right click context menu on Mac OS X

    Read the article

  • How to "restart/repair" an USB port?

    - by Click Ok
    My laptop has two USB ports, but one is broken, so I use a USB hub in the other good USB port. In that USB hub, I use a mouse and keyboard. Suddenly, that USB port doesn't detect the mouse and the keyboard (even with the light of the hub is on), and the only solution that I found is restarting the laptop. But just some minutes and the keyboard and mouse goes undetected again... Is there some method, software, etc. to "restart/repair" the USB port without restarting the PC?

    Read the article

  • Linux Bluetooth [closed]

    - by Chris
    Not sure if this is the proper forum; please forgive me if it's not... I have a Lenovo S10 'netbook' that I've installed Fedora 17 ("LXDE spin") on. So far pretty much everything works great, except, the on-board Bluetooth. lsusb shows the controller present (0a5c:2101 Broadcom Corp. Bluetooth Controller), hcitool dev shows hci0 present, but when I put my mouse ("Lenovo Bluetooth Laser Mouse," which works perfectly paired with a MacBook Pro, MacBook Air, Mac mini, and Lenovo SL500 (with a USB dongle; running Windows 7)) into pairing mode and run either hcitool scan (reports "Scanning ..." and, without further information or error message, returns to the shell prompt) or bluetooth-wizard (from the gnome-bluetooth package) and try to detect the mouse, I get nothing... Frustrating! Thanks anyone who can point me in the right direction!

    Read the article

  • Virtual/soft buttons for (Home, menu,Back, Search) always on top?

    - by Ken
    How can I make an app or maybe service that looks like (Nexus One touch buttons) for the navigation keys (Home, menu,Back, Search) The buttons should always be visibly and always stay on top and send the command to the app thats running. Someone have ideas and sample codes how to do that? * I see an app with name (Smart Taskmanager) wich always detect when you touch the right side of the screen and then detect when you slide the finger to left. So I think its possible, with this function I think its possible to implementate the code to simulate the (Home, Meny, Back, Search) buttons. * I also see and test an app wich show a "cracked display" always ontop so that tecnic maybe shold be useful to always show the buttons/bitmanp on top. Thoose function, to show the button and catch the "touch event" and send the event to the active program, thats what i dont can figure out how to do. Thats my thought! I Hope there are some deep developer wich know this solution! Regards Ken

    Read the article

  • Calling a method in a view controller from a view

    - by Lakshmie
    I have to invoke a method present in a view controller who's reference is available in the view. When I try to call the method like any other method, for some reason, iPhone just ignores the call. Can somebody explain as to why this happens and also how can I go about invoking this method? In the view I have this method: -(void) touchesBegan :(NSSet *) touches withEvent:(UIEvent *)event{ NSArray* mySubViews = [self subviews]; for (UITouch *touch in touches) { int i = 0; for(; i<[mySubViews count]; i++){ if(CGRectContainsPoint([[mySubViews objectAtIndex:i] frame], [touch locationInView:self])){ break; } } if(i<[mySubViews count]){ // viewController is the reference to the View Controller. [viewController pointToSummary:[touch locationInView:self].y]; NSLog(@"Helloooooo"); break; } } } Whenever the touches event is triggered, Hellooooo gets printed in the console but the method before that is simply ignored

    Read the article

  • server 2008 sp2 hung not responding with kvm reset as well

    - by dasko
    using server 2008 sp2 64bit os standard edition with all updates from ms site. server was hung this morning, resetting kvm did nothing, plugging in another usb mouse on the back did not let the mouse light up red on it's optic end. other machines on the kvm worked fine including the mouse. server is rack mounted 4u supermicro systems or superserver. had to hard power off and restart. any thoughts? i burnt this system in well for a couple of weeks before deploying so it is kind of odd that this happened. any help is greatly appreciated, or if anyone can suggest software to install that can maybe send out the email when something like that goes down. i looked for the minidump but nothing. nothing in the event viewers either. gd

    Read the article

  • Does anyone use the L-Track trackball?

    - by thethinman
    I've been using the Logitech Trackman Marble Mouse for years. Now I'm looking for a trackball with a scroll wheel, larger and heavier ball, and preferably rollers instead of pins. It must be finger (not thumb) operated. The Kensington Expert Mouse is close, but from what I've read the scroll wheel is poorly implemented. They also switched from rollers to pins. I bought a Kensington Orbit Trackball and it's not bad but the scroll wheel is rough and the ball is the same as the marble mouse. I'm still looking for something better. I found the L-Trac and it looks good but there's little info on the web. Has anyone used it and can provide their impressions? Or can you point out another option?

    Read the article

  • Using a KVM with a Sun Ultra 45

    - by Kit Peters
    I have a friend with two machines: a Mac and a Sun Ultra45. He wants to use a single monitor (connected via DVI) and USB keyboard / mouse with both machines. Try as I might, the Sun box will not bring up a console if the KVM is connected. However, if I connect a monitor, keyboard, and mouse directly to the Sun box, it will bring up a console that I can then (IIRC) connect to the KVM. Is there any way I can make the Sun box always bring up a console, regardless of whether it "sees" a monitor, keyboard, and mouse?

    Read the article

  • Malloc to a CGPoint Pointer throwing EXC_BAD_ACCESS when accessing

    - by kdbdallas
    I am trying to use a snippet of code from a Apple programming guide, and I am getting a EXC_BAD_ACCESS when trying to pass a pointer to a function, right after doing a malloc. (For Reference: iPhone Application Programming Guide: Event Handling - Listing 3-6) The code in question is really simple: CFMutableDictionaryRef touchBeginPoints; UITouch *touch; .... CGPoint *point = (CGPoint *)CFDictionaryGetValue(touchBeginPoints, touch); if (point == NULL) { point = (CGPoint *)malloc(sizeof(CGPoint)); CFDictionarySetValue(touchBeginPoints, touch, point); } Now when the program goes into the if statement it assigns the 'output' of malloc into the point variable/pointer. Then when it tries to pass point into the CFDictionarySetValue function it crashes the application with: Program received signal: “EXC_BAD_ACCESS”. Someone suggested not doing the malloc and pass the point var/pointer as: &point, however that still gave me a EXC_BAD_ACCESS. What I am (and it looks like Apple) doing wrong??? Thanks in advance.

    Read the article

  • Cocos2D TouchesEnded not allowing me to access sprites?

    - by maiko
    Hey Guys! Thanks so much for reading! - (void)ccTouchesEnded:(NSSet *)touches withEvent:(UIEvent *)event { UITouch * touch = [touches anyObject]; CGPoint location = [[CCDirector sharedDirector] convertToGL: [touch locationInView:touch.view]]; CGRect myRect = CGRectMake(100, 120, 75, 113); int tjx = sprite.position.x; if(CGRectContainsPoint(myRect, location)) { tjx ++; } } For some reason, ccTouchesEnded isn't allowing me to access my "sprite". I also tried to use CGRectMake like so : CGRectMake( sprite.position.x, sprite.position.y, sprite.contentSize.Width, sprite.contentSize.Height) But I couldn't access my sprites position or height. I keep getting "sprite" undeclared when it is declared in the init method, and added to the child. Please help!! I'm sure i'm missing something really simple here.

    Read the article

  • Can USB Hubs affect peripheral performance?

    - by Daniel Rasmussen
    Like many laptops, mine has 2 USB 2.0 ports. I'm looking to purchase a USB hub for my peripheral devices. I'm also planning on purchasing a laptop cooler with two USB powered fans, which connects to a USB port but can extend the plug, so it doesn't use one up. (Sorry for the poor description; see the link for a picture.) My questions are these: Can one plug 'too many' peripherals into a USB port? Can I plug my fans into a port, then the hub into the fans, then my keyboard and mouse, mic, and webcam into the extender..? Is it possible to draw too much power from a port? Secondly, will a USB hub affect the performance of any of my devices? I'm mostly worried about my mouse and keyboard. I like wired mice because I've noticed some lag in my Bluetooth mouse.

    Read the article

  • Getting x/y coordinate of a UITouch...

    - by Tarek
    HI, I have been trying to get the x/y coordinates from a touch on any iDevice. When getting the touch locations, everything looks ok if the touch is in the middle of the screen. But if I drag my finger to the bottom of the screen, I can only get a y coordinate of 1015. It should be getting to 1023. Same thing for dragging my finger to the top of the screen. I get -6. It should be 0. I have explicitly set the window and views to an origin of 0,0 and the width, height of the device's screen. Still nothing. I am really lost on what might be going on. Is something shifted? Am I not reading the x/y coordinates properly. Does something need to be transformed or converted? Any help would be much appreciated. T

    Read the article

  • How can one select an application to switch to in task switcher with track pad (MacBook Pro, Mac OS

    - by index
    Three finger swipe left/right - Task switcher shows up (a horizontal bar with app symbols). Two finger swipe - highlights an icon and shows app name. ?? - switches to highlighted app. Since the highlighting is independent of the mouse cursor, a click doesn't work here without moving the mouse cursor. But I don't want this (locate mouse cursor, move the extra mile..). Maybe Better Touch Tool allows to configure a click to act as SPACE or RETURN in the task switcher? But how did Apple intend to use this?

    Read the article

  • How to use an USB digitizer?

    - by Magnetic_dud
    Hello, i am in China now, and when i saw an USB digitizer (resistive), i bought it, thinking that i could be used as a mouse. When i connect it to my windows 7 rc1 laptop, the system properties says "pen input available" But, nothing happens when i touch the panel... i installed the bundled drivers, in chinese only, and NOT in unicode (so the text is garbled) When i install the driver, i get access to a special key on the tablet, "switch from pen input to mouse", and the driver will move the pointer according to the position on the tablet. (but no click) If i use the "pen mode", the computer will use a special chinese IME that is very awkard, i want to use the microsoft ime, and be able to draw/take notes on the screen There are "generic" drivers? Or just a plain mouse driver that will click when i press on the surface The tablet name is "qiancai"

    Read the article

  • How to zoom the desktop under Gnome?

    - by rumtscho
    I want to use my PC from the couch, but from there, I cannot see what I'm clicking. I enabled Enhanced Zoom Desktop in CCSM and it looked like it does what I want. But when "Sync mouse" is enabled, the whole desktop pans around with the tiniest mouse twitch, making me seasick. When I disable it, the mouse pointer is mapped to the non-zoomed desktop, which means that when I click on a button I am seeing on the desktop, the click registers somewhere else, and I practically cannot do anything while under zoom. Is there a better solution (does not have to involve CCSM)? Clarification: I am looking for something which zooms the entire desktop, not for a detached magnifier window.

    Read the article

  • touchesBegan doesnt get detected

    - by Muniraj
    I have a viewcontroller like the following. But the touchsBegan doestnt get detected. Can anyone plz tell me what is wrong. - (id)init { if (self = [super init]) self.view = [[[UIView alloc] initWithFrame:[[UIScreen mainScreen] applicationFrame]] autorelease]; return self; } -(void) viewWillAppear:(BOOL)animated { overlay = [[[UIImageView alloc] initWithImage:[UIImage imageNamed:@"overlay.png"]] autorelease]; [self.view addSubview:overlay]; } - (void)touchesBegan:(NSSet *)touches withEvent:(UIEvent *)event { // Detect touch anywhere UITouch *touch = [touches anyObject]; // Where is the point touched CGPoint point = [touch locationInView:self.view]; NSLog(@"pointx: %f pointy:%f", point.x, point.y); // Was a tab touched, if so, which one... if (CGRectContainsPoint(CGRectMake(1, 440, 106, 40), point)) NSLog(@"tab 1 touched"); else if (CGRectContainsPoint(CGRectMake(107, 440, 106, 40), point)) NSLog(@"tab 2 touched"); else if (CGRectContainsPoint(CGRectMake(214, 440, 106, 40), point)) NSLog(@"tab 3 touched"); }

    Read the article

  • iPhone: how do I set up a clear window-size "blocker view"?

    - by Ben
    I feel like this should be obvious to me, but for some reason I can't figure this out. I have a navigation interface with nav bar, tool bar, and primary view. Sometimes the user takes an action that causes a progress indicator to appear in the middle of the view. While the progress indicator (which is a custom UIView) in spinning in the middle, I want no touch input to be allowed to go to any of the underlying interface (main view, nav bar, toolbar, etc). But this doesn't seem trivial. I've tried (and failed) to create a simple view whose only job is to swallow touch input and use it as a window subview-- no dice, it never gets the touch events (and yes, it does have userInteractionEnabled). I've tried to bolt it on as a transparent modal view controller, but those don't seem to ever be transparent. Thoughts? What am I missing? Thanks!

    Read the article

  • Debian Squeeze: Trackpad on EEE PC 1000HA acts like slate

    - by CarlF
    I'm using Debian on an Asus Eee PC 1000HA. It works flawlessly, except that the built-in trackpad acts as if it was a slate. That is, if I touch the upper-left quadrant of the trackpad, the mouse pointer "teleports" to the upper left of the netbook's screen. If I touch the bottom center, the pointer teleports to the bottom center of the screen. I currently have no xorg.conf. I have tried various settings for the mouse in xorg.confs that I created, and none of them did more than disable the trackpad entirely. The trackpad works perfectly in Windows XP (shipped with the Eee) and in Ubuntu. Debian works fine with the (USB) mouse. Any suggestions? Thanks.

    Read the article

< Previous Page | 142 143 144 145 146 147 148 149 150 151 152 153  | Next Page >