Search Results

Search found 9614 results on 385 pages for 'touch mouse'.

Page 134/385 | < Previous Page | 130 131 132 133 134 135 136 137 138 139 140 141  | Next Page >

  • How to extend selection in gnome-terminal?

    - by tomasorti
    In a terminal I can select a single line, double clicking with the Left Mouse Button. With xterm, I can extend that selection clicking with the Right Mouse Button in the place I want to extend it. Then can I paste the whole selection with the Middle Mouse Button or paste it in other application. In gnome-terminal, it seems I can extend the selection clicking with the Left Mouse Button, but holding at the same time the Shift key. Visually, it seems the selection is done, but when clicking with the Middle Mouse button or pasting in other application, I only get the primary single line selection. How can I get the whole selection under gnome-terminal? Is it possible to use selections in gnome-terminal as xterm does? Cheers, Tomas.

    Read the article

  • iPod touch debugging: Error on install/run only if app exists on device already?

    - by Ben
    Hi all, I am using an iPod to test an app. The device is all set up with the right provisioning profiles, etc-- that's not really the issue. But every time I start the app from Xcode on the device, I get the "A signed resource has been added, modified, or deleted." error from the Organizer window. Wait, I know, you think it's a provisioning profile problem. But here's the kicker: if I just delete the app from the iPod (using the main screen) and try again, it works fine. I only get this error when the app is already installed. The other kicker is that this behavior doesn't happen on an iPhone that I have for occasional testing-- on that device, I can start/restart/restart indefinitely. But using the iPod, my compile-run-test cycle is annoyingly slow since I have to manually delete the app each time. Any ideas? I'm using Xcode 3.2.2 (prerelease) FWIW. The iPod has stock OS 3.1.2 on it. Thanks!

    Read the article

  • Invisible boundary on Synergy client

    - by Jayson Rowe
    I have a Desktop system (named 'desktop') with 2 monitors attached (one 1680x1050 and another 1920x1080 resolution), running a synergy server. I also have a small ITX machine (named 'tiny') to my right that has a single monitor running at 1280x1024. The dual-monitor desktop is the synergy "server" and the single monitor system is the "client". Synergy works fine with one exception; if I move the mouse to the client, I can't move my mouse below the top two-thirds of the screen - it just stops. I can grab a mouse attached to the computer, and it moves all the way down, but as soon as I touch the mouse attached to the server, it jumps back to the top of the screen. Is there an issue with my config? section: screens desktop: tiny: end section: links desktop: right = tiny tiny: left = desktop end Thanks in advance for any suggestions. +----------------------------+-----------------------------+ | | | | | | | | +-----------------+ | | | | | desktop 1680x1050 | desktop 1920x1080 | | | | | | | | | tiny 1280x1024 | | | |+---------------+| | | |XXXXXXXXXXXXXXXXX| +----------------------------+-----------------------------+-----------------+

    Read the article

  • Which is a better option, OpenGL or a game engine for developing a game for the iphone or ipod touch

    - by balraj
    I am new to OpenGL ES, and I'm about to begin a 3D game for the iphone in which we are showing some car pursuit or racing. Is it possible just with the OpenGL ES or UIKit only, or do I have to use other tools for it? I am comfortable with UIKit but newer to OpenGL/OpenGL ES; which would be better to start this game? Or should I use a game engine? If so, then which game engine would give us the 3D feeling, quality of images and motion, and rendering of the views with the sound effects?

    Read the article

  • how to detect touch on UIView with UIimage ontop of it?

    - by Silent
    hello all, i want to use a UIview and a UIImage ontop of it with a text box and button i want the user to drag it when the UIview or anything in the view is touched, what would be the best way to implement this method? the UIView is on a single View Controller and using Interface Builder to make my UIviews and UIimages.

    Read the article

  • How to disable other touch gestures after adding another tap gesture to the view?

    - by Hudson Duan
    I have a view with some tables and buttons on it, and then I want to add a tap gesture to the entire view, but I only want that gesture recognizer to recognize taps. Ideally, I want to do something when the added gesture recognizer is tapped, then remove that gesture recognizer after so the other buttons and tables can be accessed. Basically a tap to dismiss functionality that replicates something like the facebook notifications window, tap outside to dismiss, but not interfere with the buttons outside of the notifications view. Can anybody help? My current code is: NotificationsWindow *customView = [[[NSBundle mainBundle]loadNibNamed:@"NotificationsWindow" owner:self options:nil]objectAtIndex:0]; customView.frame= CGRectMake(12, 12, customView.frame.size.width, customView.frame.size.height); UITapGestureRecognizer *recognizerForSubView = [[UITapGestureRecognizer alloc] initWithTarget:self action:@selector(handleTapBehindAgain:)]; [recognizerForSubView setNumberOfTapsRequired:1]; recognizerForSubView.cancelsTouchesInView = NO; //So the user can still interact with controls in the modal view [customView addGestureRecognizer:recognizerForSubView]; [self.view addSubview:customView]; UITapGestureRecognizer *recognizerForSuperView = [[UITapGestureRecognizer alloc] initWithTarget:self action:@selector(handleTapBehind:)]; [recognizerForSuperView setNumberOfTapsRequired:1]; recognizerForSuperView.cancelsTouchesInView = NO; //So the user can still interact with controls in the modal view [customView.superview addGestureRecognizer:recognizerForSuperView]; (void)handleTapBehind:(UITapGestureRecognizer *)sender { NSLog(@"tapped"); [[self.view.subviews lastObject] removeFromSuperview]; [self.view removeGestureRecognizer:sender]; } I want to make it so that the recognizer for the super view dismisses the subview, but not to interfere with the other taps on the super view.

    Read the article

  • Ubuntu 12.04 Precise on Dell Inspirion Duo

    - by Roman M. Kos
    I installed 12.04 version of Ubuntu on my Dell Inspiron Duo. After installing with help of program like UNetBootIn or smth like that. ( Besides i have no problem with kernel panic on chargin on/off like in 11.10. ) After that i followed with steps from here, in the first post here: http://ubuntuforums.org/showthread.php?t=1658635 There left one big problem with touchscreen: When i whant to drag with touchscreen like i`m doing it with mouse (for ex. selecting multile files with mouse) the selectable rectangular doesnot shows while im dragging, when dragging was finished (i put my finger off) it shows the rectangular and hides it. This thing disables all my tries to drag a window or smth else.... Also some time after using touchscreen such things are disabeling: - Often - click from a mouse (after keyboard using functinality restores) - less often - mouse movement disables (sometimes restores sometimes not) - lesser than other - keyboards works but no sygnals accepting (keyoards has indicator, so thay react, mouse of course not) The test from eTouchU utility passes perfeclty. Any idia for solving this problem? P.S.: Im from Ukraine, so sorry if my possible grammar mistakes. P.P.S.: Besides how to know the physicall position of my tabled mode? For automaticall rotating. Like on each rotation do some script.

    Read the article

  • Jittery Movement, Uncontrollably Rotating + Front of Sprite?

    - by Vipar
    So I've been looking around to try and figure out how I make my sprite face my mouse. So far the sprite moves to where my mouse is by some vector math. Now I'd like it to rotate and face the mouse as it moves. From what I've found this calculation seems to be what keeps reappearing: Sprite Rotation = Atan2(Direction Vectors Y Position, Direction Vectors X Position) I express it like so: sp.Rotation = (float)Math.Atan2(directionV.Y, directionV.X); If I just go with the above, the sprite seems to jitter left and right ever so slightly but never rotate out of that position. Seeing as Atan2 returns the rotation in radians I found another piece of calculation to add to the above which turns it into degrees: sp.Rotation = (float)Math.Atan2(directionV.Y, directionV.X) * 180 / PI; Now the sprite rotates. Problem is that it spins uncontrollably the closer it comes to the mouse. One of the problems with the above calculation is that it assumes that +y goes up rather than down on the screen. As I recorded in these two videos, the first part is the slightly jittery movement (A lot more visible when not recording) and then with the added rotation: Jittery Movement So my questions are: How do I fix that weird Jittery movement when the sprite stands still? Some have suggested to make some kind of "snap" where I set the position of the sprite directly to the mouse position when it's really close. But no matter what I do the snapping is noticeable. How do I make the sprite stop spinning uncontrollably? Is it possible to simply define the front of the sprite and use that to make it "face" the right way?

    Read the article

  • How to deal with forward declaration / #import in Cocoa Touch (Objective-C cross C++) correctly?

    - by unknownthreat
    I am trying to write this header file: //@class AQPlayer; //#import "AQPlayer.h" @interface AQ_PWN_iPhoneViewController : UIViewController { AQPlayer* player; } @end AQPlayer is a .mm file written in C++. I tried to make a class forward declaration here, but it complains to me: error: cannot find interface declaration for 'AQPlayer' So I tried to "#import" the header file instead, but it complains something completely off and weird. Here's a slice of the error complained: In file included from /Users/akaraphan/Desktop/SpecialTopic1/AQ_PWN_iPhone/Classes/AQPlayer.h:51, from /Users/akaraphan/Desktop/SpecialTopic1/AQ_PWN_iPhone/Classes/AQ_PWN_iPhoneViewController.h:12, from /Users/akaraphan/Desktop/SpecialTopic1/AQ_PWN_iPhone/Classes/AQ_PWN_iPhoneAppDelegate.m:10: /Developer/Examples/CoreAudio/PublicUtility/CAStreamBasicDescription.h:78: error: expected '=', ',', ';', 'asm' or '__attribute__' before 'CAStreamBasicDescription' /Developer/Examples/CoreAudio/PublicUtility/CAStreamBasicDescription.h:230: error: expected '=', ',', ';', 'asm' or '__attribute__' before '<' token /Developer/Examples/CoreAudio/PublicUtility/CAStreamBasicDescription.h:231: error: expected '=', ',', ';', 'asm' or '__attribute__' before '==' token /Developer/Examples/CoreAudio/PublicUtility/CAStreamBasicDescription.h:233: error: expected '=', ',', ';', 'asm' or '__attribute__' before '!=' token /Developer/Examples/CoreAudio/PublicUtility/CAStreamBasicDescription.h:234: error: expected '=', ',', ';', 'asm' or '__attribute__' before '<=' token /Developer/Examples/CoreAudio/PublicUtility/CAStreamBasicDescription.h:235: error: expected '=', ',', ';', 'asm' or '__attribute__' before '>=' token /Developer/Examples/CoreAudio/PublicUtility/CAStreamBasicDescription.h:236: error: expected '=', ',', ';', 'asm' or '__attribute__' before '>' token /Developer/Examples/CoreAudio/PublicUtility/CAStreamBasicDescription.h:239: error: expected ';', ',' or ')' before '&' token Am I missing something? Can't I do a forward declaration for this?

    Read the article

  • How do Unity 12.04/Compiz bindings really work?

    - by Daniel
    There is a bewildering array of places to set bindings, all inconsistent with one another. E.g. in Unity's System Settings having the Ctrl-key highlight the mouse position is an on/off choice. I like the feature, but not on such a prominent key where I keep activating it accidentally. The keyboard shortcuts allow only one binding per command, where I might like a convenient one on the external keyboard and an emergency alternative for when I'm on the road. Keyboard custom shortcuts has a nice interface, but allows only key bindings — besides it doesn't seem to work for me. So I activated CCSM Commands. There I have the choice of key-, mouse- and/or edge bindings. Whereas some places in CCSM offers only one or two of these binding possibilities, randomly at the whim of the programmer. I have not found a way to differentiate a mouse-drag from a click. E.g. I want <SuperMouse1-drag anywhere on a window to move it, while if I don't drag, it should be raise-lower. On the title bar I want the same without needing the <Super key. Now I find raise-lower only in System Settings where I can't assign a mouse binding. If therefore in CCSM I fallback to only lower and put move on the same binding, the window already gets lowered on mouse down, and I can then invisibly move it. Very useful! I have <Altasciicircum get in the way of an Emacs binding, with some to me useless popup overlay. I can find it nowhere, so I can't turn it off. So how can I go without these frontends until they have matured, and instruct Compiz directly, for example in the way Emacs or Sawfish have keymaps, and separate ones for each context, with inheritance?

    Read the article

  • jquery ondrag map load only what has not been viewed

    - by David
    When a person mouse down, moves the mouse, and mouses up the system gets the different in the mouse down coord and the mouse up coords and loads in the new map items. However, the problem is it loads them every time so I want a way to track what has been loaded without too much work on checking or storing checks. Most promising. Came up with while typing this. Section the screen into 250x250 sectors and check if that sector has been loaded. Keep track of each corner of the screen and see if there is an area of those that have not loaded. Keep a record of the corners of the screen. When mouse up coords are greater then load the different. Problem is if they are at coord 10,000 then it will load from -10k to 10k positive and that is a lot of items to load. Check every item on the page to see if it is loaded. If I do this I might as well reload the whole page. If anyone has some suggestions; feel free to pass them on.

    Read the article

  • Introducing the Metro User Interface on Windows 2012

    - by andywe
    Although I am a big fan of using PowerShell to do many of my server operations, that aspect is well covered by those far more knowledgeable than I, and there is vast information around the web already on that. The new Metro interface, and getting around both Windows 8 and Windows Server 2012 though is relatively new, even for those whop ran the previews. What is this? A blank Desktop!   Where did the start button go? Well, it is still there...sort of. It is hidden, and acts like an auto hidden component that appear only when the mouse is hovered over the lower left corner of the screen. Those familiar with Gnome or OSX can relate this to the "Hot Corners" functions. To get to the start button, hover your mouse in the very left corner of the task bar. Let it sit there a moment, and a small blue square with colored tiles in it called start will appear. Click it. I clicked it and now I have all the tiles..What is this?   Welcome to the Metro interface. This is a much more modern look, and although at first seems weird and cumbersome, I have actually found that it is a bit more extensible, allowing greater organization and customization than the older explorer desktop. If you look closely, you'll see each box represents either a program, or program group. First, a few basics about using the start view. First and foremost, a right mouse click will bring up a bar on the bottom, with an icon towards the right. Notice it is titled “All Apps”. An even easier way in many places is to hover your mouse in the exact opposite corner, in the upper right. A sidebar will open and expose what used to be a widget bar (remember Vista?), and there are options for Search, Start, and Settings.   Ok Great, but where is everything? It’s all there…Click the All Apps icon.   Look better? Notice the scroll bar at the bottom. Move it right..your desktop is sized to your content..so you can have a smaller, or larger amount of programs exposed. Each icon can be secondary clicked (right mouse click for most of us, and an options bar at the bottom, rather than the old small context menu, is opened with some very familiar options.   Notice the top of the Windows Explorer window has some new features. You still have your right mouse click functions, but since the shortcuts for these items already exist..just copy them. There are many ways, but here is a long way to show you more of the interface. 1. Right mouse click a program icon, and select the Open File Location option. 2. Trusty file manager opens…but if you look closely up at top edge of the window, you’ll see a nifty enhancement. An orange colored box that is titled Shortcut Tools and another lavender box Title Application tools. Each of these adds options at the top of the file manager window to make selection easy. Of course, you can still secondary click an item in the listing window too. 3. Click shortcut tools, right click your app shortcut and copy it. Then simply paste it into the desktop outside the File Explorer window Also note some of the newer features. The large icons up top below the menu that has many common operations. The options change as you select each menu item. Well, that’s it for this installment. I hope this helps you out.

    Read the article

  • Android browser touch events stop display being updated inc. canvas/elements - How to work around?

    - by Ed Kirk
    On some android's native browser touching the page seems to stop the display from being updated until the finger is released. This occurs for both html element based animation (switching classes) and for canvas based animation. It does not however stop normal js execution and other events are fired as normal. On devices with this problem the dolphin browser also seems effected (not firefox though). Touchstart/move both have preventDefault() fired as well as stopPropergation(), cancelBubble = true; and e.returnValue = false;. In the CSS webkit selection has also been disabled. The page will not scroll. A similar question has been asked here: Does Android browser lock DOM on touchStart? but I'd like to find out if this behaviour can be overcome, or at least to discover what devices will be effected by the problem, is it a device or version android issue? If you cannot answer the question running the demo and reporting your experience along with your device model and useragent (displayed at bottom of demo page) as a comment might help others or myself answer the question. Here is a demo and steps to reproduce the behaviour. A QR code for the link can be found here https://s3-eu-west-1.amazonaws.com/canvas-test-pd/tmp.png. https://s3-eu-west-1.amazonaws.com/canvas-test-pd/index.html The web page has a canvas at the top and a div with a background image at the bottom. Every second the canvas is cleared and a different image displayed and the div has it's class switched (both toggle between 0 and 1 pngs). Once this has toggled a few times place your finger on the canvas (the top grey box) and hold it there. Wait to see if the animation continues (sometimes it will once or twice then stops) and if there are any visual distortions. Update It seems that the Galaxy Tab running 3.2 requires handlers for touchstart/end of document, not just required divs for the screen to continue updating the display. Thanks jimpic. I'm starting to believe it's an issue caused by manufacturers skins, although this is difficult to prove.

    Read the article

  • Safari on the iPhone & iPad gives colour feedback on touch, I want to stop this.

    - by Owen
    Clicking on an element which has a Javascript handler makes the element go have a 'grey overlay'. This is normally fine but I'm using event delegation to handle the touchdown events of many child elements. Because of the delegation the 'grey overlay' is appearing over the parent element and looks bad and confusing. I could attach event handlers to the individual elements to avoid the problem but this would be computationally very wasteful. I'd rather have some webkit css property that I can override to turn it off. I already have visual feedback in my app so the 'grey overlay' is not needed. Any ideas?

    Read the article

  • How to detect the active iTunes store on the iPhone/iPod Touch/iPad?

    - by Paul
    I'd like to be able to determine which store the user connects to from inside my app, so that I can direct them to some appropriate content for their device AND store. Does anyone know how to get this information? Basically, if the user is in the UK, and connects to the UK store, I want my function/method to return GB, if in Korea, I want KR, Australia = AU etc. Any help would be appreciated.

    Read the article

  • TabCtrl

    Adjustable contol with zooming and scrolling tabs, dragging with the mouse, custom drawing and much more

    Read the article

  • Is two-finger non-homerow touch-typing for programming acceptable?

    - by codebliss
    I'm currently typing about 90 wpm (from http://speedtest.10-fast-fingers.com/ 90 correct 0 missed) using two fingers and the occasional ring or index. This probably grew from learning to type at an early age, before home-row was presented to me. Is this acceptable? Do people religiously endorse home-row even with low-mistake poking without looking at the keyboard?

    Read the article

  • How can I get touch co-ordinates in cocos2d?

    - by Nouman
    Hi, I am a newbee in iPhone game development. What I want is that wherever user touches on iPhone screen, where I will display a pic at that place. Can anyone help me how to get co-ordinates of the area everytime when user touches the screen? Regards, Nouman

    Read the article

  • "There was an internal API error." while running an app on any iPhone/iPod-touch device

    - by Martin Cowie
    I am in the process of submitting an iPhone app to the app store. While making the final touches to the app I was in the process of compiling and running the app on my iPhone when I got the message ... "There was an internal API error." The console had this to say ... 25/08/2010 10:10:54 Xcode[3556] Failed willExecute: Error Domain=com.apple.platform.iphoneos Code=0 UserInfo=0x2011adec0 "There was an internal API error." -- { NSLocalizedDescription = "There was an internal API error."; NSLocalizedFailureReason = ""; NSLocalizedRecoverySuggestion = ""; } The problem is specific to this project, others projects don't suffer the same problem. The same problem exhibits when moved to another machine, or another mobile device is swapped in. I should be most grateful for any hints or ideas on the subject ...

    Read the article

< Previous Page | 130 131 132 133 134 135 136 137 138 139 140 141  | Next Page >