Search Results

Search found 41766 results on 1671 pages for 'two finger drag'.

Page 182/1671 | < Previous Page | 178 179 180 181 182 183 184 185 186 187 188 189  | Next Page >

  • MySQL Multi-Aggregated Rows in Crosstab Queries

    MySQL's crosstabs contain aggregate functions on two or more fields, presented in a tabular format. In a multi-aggregate crosstab query, two different functions can be applied to the same field or the same function can be applied to multiple fields on the same (row or column) axis. Rob Gravelle shows you how to apply two different functions to the same field in order to create grouping levels in the row axis.

    Read the article

  • MySQL Multi-Aggregated Rows in Crosstab Queries

    MySQL's crosstabs contain aggregate functions on two or more fields, presented in a tabular format. In a multi-aggregate crosstab query, two different functions can be applied to the same field or the same function can be applied to multiple fields on the same (row or column) axis. Rob Gravelle shows you how to apply two different functions to the same field in order to create grouping levels in the row axis.

    Read the article

  • Gmail Rolls Out New Compose Features

    - by Jason Fitzpatrick
    Gmail has several new features that make it even easier to compose email messages including pop-over compositions windows (similar to the Google Chat window), contact profile pictures in the address box, and drag and drop address switching. If you’ve ever had to open two separate windows in order to continually tab back and forth so you could reference one email while composing another, you’ll certainly appreciate the new pop-over compose window that allows you to work within Gmail while keeping a small email composition window open in the corner–as seen in the screenshot above. In addition to that major change, Gmail has also introduced contact photos in the address suggestion window (making it easier than ever to make sure you’re selecting the right recipient) and the ability to drag and drop addresses between the To:, CC:, and BCC: address slots. Introducing the New Compose in Gmail [The Official Gmail Blog] 6 Start Menu Replacements for Windows 8 What Is the Purpose of the “Do Not Cover This Hole” Hole on Hard Drives? How To Log Into The Desktop, Add a Start Menu, and Disable Hot Corners in Windows 8

    Read the article

  • Why can I only put some items in the launcher?

    - by HaskellElephant
    Whenever I search for an item using dash it knows the default way of opening it. However if I drag a file into the launcher it can't be added, even if it is executable. For example if you would like minecraft to show up you will have to add a custom launcher, or if you use dash to find a pdf file, you can double click it and it opens evince however you cannot drag it into the launcher. The animation surrounding this behavior suggests that this was done by design. Why is this useful or necessary? Is it a bug?

    Read the article

  • How can one select an application to switch to in task switcher with track pad (MacBook Pro, Mac OS

    - by index
    Three finger swipe left/right - Task switcher shows up (a horizontal bar with app symbols). Two finger swipe - highlights an icon and shows app name. ?? - switches to highlighted app. Since the highlighting is independent of the mouse cursor, a click doesn't work here without moving the mouse cursor. But I don't want this (locate mouse cursor, move the extra mile..). Maybe Better Touch Tool allows to configure a click to act as SPACE or RETURN in the task switcher? But how did Apple intend to use this?

    Read the article

  • Best way to prevent UIPanGestureRecognizer from firing when moving sprites in cocos2d

    - by cjroebuck
    Im using UIPanGestureRecognizer in my cocos2d game to do drag and drop of sprites. I have a row of sprites and when I drag a sprite on top of another one, the sprite underneath it and any other sprites between should shift left or right out of the way to allow space to drop the currently selected sprite. This is working ok, however, if I am too quick at dragging the sprite around the screen, this triggers another round of the UIPanGestureRecognizer's callback method, and screws up the logic, as the sprites are in-between shifting. I need a way to freeze the callback from firing, whilst the other sprites are shifting, then once they have finished moving, re-enable the callback to fire. Whats the best way to do this?

    Read the article

  • Platinum SEO Plugin Vs All-In-One SEO Pack

    The two biggest SEO optimization tools for WordPress are probably Platinum SEO and All-in-One SEO Pack. This article explains the relationship of the two and gives a brief rundown of installing Platinum SEO plugin on your blog. It then provides a comparison of the two, feature-for-feature.

    Read the article

  • How to set up Mac Os X like dragging behaviour on Linux

    - by ticking
    I would like to use Linux(Fedora) on a Aluminum MacBook Pro, but since the Touchpad on a MacBook Pro is only a giant button Apple does some custom tracking. When there is one finger placed, and a second follows (the click can occur before or after that) it will be interpreted as a drag. So the strong thumb can be used to hold pressure and the more accurate index finger can do the pointing. But Linux interprets this as a right click, since it only cares if two fingers are on the pad. Is there a way to achieve said behaviour? Cheers Jan

    Read the article

  • Why doesn't the "scroll bar" on my laptop touchpad work?

    - by Antonio Guzman Belmont
    EDIT, from Google translate: Well buy my laptop with windows 7 and the touchpad has some arrows up and down to q deslises finger on the right side as if the q scroll the scroll bar handles, recently installed Ubuntu 11.10 and there is also served by the touchpad scroll so to speak but then asked me and I did update (late hours) and longer term attempt q dezlizar my finger down the right side of the touchpad and now low, anyone can tell me I can do about q to q and windows function as and ubuntu 11.10, I think now I have ubuntu 4.12 EDIT.2 (minimal tidy-up) I bought my laptop with Windows 7 installed. The touchpad has up/down arrows on the right side, as if this would scroll. When I installed Ubuntu 11.10, I was able to use this touchpad scroll function. I recently did an update (at some late hours) and i can no longer use right side of the touchpad to scroll. Can anyone tell me how to get the same function as I had in Windows and Ubuntu 11.10 ? I think I now have Ubuntu 12.04 Original question: Bueno compre mi lap con windows 7 y en el touchpad tiene unas flechitas arriba y abajo para q deslises el dedo del lado derecho como si fuese el scroll q maneja la barra de desplazamiento, recientemente instale ubuntu 11.10 y ahi servia tambien el scroll del touchpad por asi decirlo pero luego me pidio actualizar y lo hice(tardo horas) y ya q termino intento dezlizar mi dedo para bajar por la parte derecha del touchpad y ahora no baja, alguien me puede decir q puedo hacer al respecto para q funcione como e windows y en ubuntu 11.10, ahora tengo ubuntu 12.04 creo

    Read the article

  • ~/.xinput.d folder is ignored in Ubuntu 13.04

    - by CaptSaltyJack
    It used to be that you could make a file ~/.xinput.d/en_US and put xinput commands in there, such as enabling drag lock. Now, for some reason, in 13.04 this does not work. Anyone know why this changed, and how to set these? I suppose I could just put the xinput commands in a script file and have it execute upon login. I'm just wondering why the old method stopped working. EDIT: Current file /etc/X11/xinit/xinput.d/en_US: xinput set-prop 17 316 1 xinput set-prop 17 317 350 But I've realized that for some reason, the touchpad ID changes. Right now it's 15. Also, the actual properties such as "Drag Lock" can change. So this method doesn't work.

    Read the article

  • Cocos2d Tiled Dynamic Object Layer

    - by Rodrigo Camargo
    I'm trying to develop a cocos2d tiled based game using a sort of 'dynamic' object layer. What I want to do is after the tiled map is loaded, the user can drag something into the map and that will become an event when the 'hero' pass over it. I know how to build an object layer in tiled but it seems that is for fixed positions and what I want is a dynamic action position based on what the user can select. For instance, the user can drag a rock into a tile and when the character hit that rock he may die, or something. I'm a little lost about how to make it work. Do you have any idea of what should I use or what should I look for? Thanks in advance!

    Read the article

  • How do I fix dragging windows to the adjacent workspace?

    - by Bill O'Dwyer
    I installed CompizConfig-Settings-Manager and I put on all the settings I liked and had in 11.10, including the ability to drag my windows to the adjacent workspace. It's under the Desktop Wall section, on the Edge Flipping tab and I've checked "Edge Flip Move" and "Edge Flip DnD." In 11.10, the movement was smooth between each workspace, and the window would still be "grabbed" in the same place. In 12.04, it's leaving the window behind and the mouse appears to be "grabbing" nothing, but I'm still holding onto the window, and I can still move and place it within the workspace (or indeed the previous workspace as it won't appear in the desired place until I drag the mouse all the way to the edge of the screen). Any way to fix this? I'm running 12.04 beta 2.

    Read the article

  • Zoom Layer centered on a Sprite

    - by clops
    I am in process of developing a small game where a space-ship travels through a layer (doh!), in some situations the spaceship comes close to an enemy space ship, and the whole layer is zoomed in on the two with the zoom level being dependent on the distance between the ship and the enemy. All of this works fine. The main question, however, is how do I keep the zoom being centered on the center point between the two space-ships and make sure that the two are not off-screen? Currently I control the zooming in the GameLayer object through the update method, here is the code (there is no layer repositioning here yet): -(void) prepareLayerZoomBetweenSpaceship{ CGPoint mainSpaceShipPosition = [mainSpaceShip position]; CGPoint enemySpaceShipPosition = [enemySpaceShip position]; float distance = powf(mainSpaceShipPosition.x - enemySpaceShipPosition.x, 2) + powf(mainSpaceShipPosition.y - enemySpaceShipPosition.y,2); distance = sqrtf(distance); /* Distance > 250 --> no zoom Distance < 100 --> maximum zoom */ float myZoomLevel = 0.5f; if(distance < 100){ //maximum zoom in myZoomLevel = 1.0f; }else if(distance > 250){ myZoomLevel = 0.5f; }else{ myZoomLevel = 1.0f - (distance-100)*0.0033f; } [self zoomTo:myZoomLevel]; } -(void) zoomTo:(float)zoom { if(zoom > 1){ zoom = 1; } // Set the scale. if(self.scale != zoom){ self.scale = zoom; } } Basically my question is: How do I zoom the layer and center it exactly between the two ships? I guess this is like a pinch zoom with two fingers!

    Read the article

  • Upgrading your Internet connection to obselescence

    - by TATWORTH
    Originally posted on: http://geekswithblogs.net/TATWORTH/archive/2013/10/11/upgrading-your-internet-connection-to-obselescence.aspxRecently I was approached by two UK Internet Service Providers to upgrade to a fibre connection. In each case I asked two questions:Can you provide me with a fixed IP Address?Is the equipment to be provided compatible with IPv6?None of the question was satisfactorily answered. One of the persons answering even suggested that granting a fixed IP Address would be a breach of security!I find it very disturbing that two companies that present themselves as innovative should still not be preparing for IPv6. The answer I would have expected was that all new equipment being supplied was IPv6 compatible and that plans were in hand for a switchover to IPv6. Instead new equipment would be supplied that would have to be replaced when IPv6 comes. Equally disturbing was that the call center people who answered did not know why a fixed IP address was important or why the change to IPv6 would have to come.I would rather not name and shame the two companies, however I will be looking elsewhere for my next Internet Service Provider.

    Read the article

  • Can I still right click on an app to move it to the left or right workspace like in 11.04 and before? What about one click switcher?

    - by Thierry Lam
    In 11.04(classic edition) and before, one of the best features of Ubuntu was the workspace switcher, none of the other major OS had something like that built natively. I work with 2 workspaces, one left and the other right. I will right click apps from left workspace and throw them to the right and vice versa. In 11.10, I have to: Open workspace from the launcher Drag the app to the other workspace Double click back on my original workspace What used to take 2 clicks, now takes 1 click, 1 double click and a drag. Is that the modern way to switch apps between workspaces? In older Ubuntu, I had the 2 workspace icons at the bottom right corner of my screen. I can then single click on the left or right workspace for switching. I'm now doing step 1 and 3 from above to perform that single. Are there any shortcuts or customization to reduce the number of events I have to go through to perform those simple actions?

    Read the article

  • My first flash game bot, in java

    - by Dylan
    Okay so i love coming up with new programming challenges and ive discovered a new challenge. I would love to create a bot for a game that requires the user to click on a character and drag the mouse like a slingshot. Upon releasing the mouse the character flys across the game and hopefully lands in a scored spot(in my bot the highest score). the game looks like this an image of the game is here. http://i.stack.imgur.com/fThnG.jpg How would i go about calculating the location of the character and then the physics to know exactly where to drag the mouse to?

    Read the article

  • How do I disable window move with alt + left mouse button in GNOME Shell?

    - by artfulrobot
    There are lots of posts on this about Unity, this is about gnome-shell As with other posters, I use apps (Inkscape in particular) that uses Alt-Click and Alt-drag. Gnome-shell is grabbing this off me and thus disabling features in the application. I have tried using ccsm's Move plugin - you can turn this off and you still get Alt-drag window moving. using gconf editor to change /apps/metacity/general/mouse_button_modifier - this has no effect AFAICS. It's really hindering me. Otherwise I find gnome-shell a really productive environment.

    Read the article

  • How to prevent a Gtk.SourceView from grabbing the URI of a file

    - by laboleite
    I'm making a text editor using GTK3 in Vala. I need to drag a file and, when I drop it into the application window, the application opens that file. I'm using a Gtk.Notebook that contains the Gtk.SourceView's. It works when I drop a file into an empty Gtk.Notebook, but when there is at least one Gtk.SourceView's appended on it, the SourceView grab the URI of the file, shows it and the window can't handle the opening of the file. In this case, what can I do to prevent a Gtk.SourceView from grabbing the uri of a file when I drop a file into a Gtk.SourceView. PS: I tried to use the drag_dest_unset() inside the SourceView derived class. It worked, the SourceView didn't grab the URI and the window could open the file, but the application showed a message like this, in runtime: Gtk-WARNING **: Can't set a target list on a widget until you've called gtk_drag_dest_set() to make the widget into a drag destination

    Read the article

  • Navigation in Win8 Metro Style applications

    - by Dennis Vroegop
    In Windows 8, Touch is, as they say, a first class citizen. Now, to be honest: they also said that in Windows 7. However in Win8 this is actually true. Applications are meant to be used by touch. Yes, you can still use mouse, keyboard and pen and your apps should take that into account but touch is where you should focus on initially. Will all users have touch enabled devices? No, not in the first place. I don’t think touchscreens will be on every device sold next year. But in 5 years? Who knows? Don’t forget: if your app is successful it will be around for a long time and by that time touchscreens will be everywhere. Another reason to embrace touch is that it’s easier to develop a touch-oriented app and then to make sure that keyboard, nouse and pen work as doing it the other way around. Porting a mouse-based application to a touch based application almost never works. The reverse gives you much more chances for success. That being said, there are some things that you need to think about. Most people have more than one finger, while most users only use one mouse at the time. Still, most touch-developers translate their mouse-knowledge to the touch and think they did a good job. Martin Tirion from Microsoft said that since Touch is a new language people face the same challenges they do when learning a new real spoken language. The first thing people try when learning a new language is simply replace the words in their native language to the newly learned words. At first they don’t care about grammar. To a native speaker of that other language this sounds all wrong but they still will be able to understand what the intention was. If you don’t believe me: try Google translate to translate something for you from your language to another and then back and see what happens. The same thing happens with Touch. Most developers translate a mouse-click into a tap-event and think they’re done. Well matey, you’re not done. Not by far. There are things you can do with a mouse that you cannot do with touch. Think hover. A mouse has the ability to ‘slide’ over UI elements. Touch doesn’t (I know: with Pen you can do this but I’m talking about actual fingers here). A touch is either there or it isn’t. And right-click? Forget about it. A click is a click.  Yes, you have more than one finger but the machine doesn’t know which finger you use… The other way around is also true. Like I said: most users only have one mouse but they are likely to have more than one finger. So how do we take that into account? Thinking about this is really worth the time: you might come up with some surprisingly good ideas! Still: don’t forget that not every user has touch-enabled hardware so make sure your app is useable for both groups. Keep this in mind: we’re going to need it later on! Now. Apps should be easy to use. You don’t want your user to read through pages and pages of documentation before they can use the app. Imagine that spotter next to an airfield suddenly seeing a prototype of a Concorde 2 landing on the nearby runway. He probably wants to enter that information in our app NOW and not after he’s taken a 3 day course. Even if he still has to download the app, install it for the first time and then run it he should be on his way immediately. At least, fast enough to note down the details of that unique, rare and possibly exciting sighting he just did. So.. How do we do this? Well, I am not talking about games here. Games are in a league of their own. They fall outside the scope of the apps I am describing. But all the others can roughly be characterized as being one of two flavors: the navigation is either flat or hierarchical. That’s it. And if it’s hierarchical it’s no more than three levels deep. Not more. Your users will get lost otherwise and we don’t want that. Flat is simple. Just imagine we have one screen that is as high as our physical screen is and as wide as you need it to be. Don’t worry if it doesn’t fit on the screen: people can scroll to the right and left. Don’t combine up/down and left/right scrolling: it’s confusing. Next to that, since most users will hold their device in landscape mode it’s very natural to scroll horizontal. So let’s use that when we have a flat model. The same applies to the hierarchical model. Try to have at most three levels. If you need more space, find a way to group the items in such a way that you can fit it in three, very wide lanes. At the highest level we have the so called hub level. This is the entry point of the app and as such it should give the user an immediate feeling of what the app is all about. If your app has categories if items then you might show these categories here. And while you’re at it: also show 2 or 3 of the items itself here to give the user a taste of what lies beneath. If the user selects a category you go to the section part. Here you show several sections (again, go as wide as you need) with again some detail examples. After that: the details layer shows each item. By giving some samples of the underlaying layer you achieve several things: you make the layer attractive by showing several different things, you show some highlights so the user sees actual content and you provide a shortcut to the layers underneath. The image below is borrowed from the http://design.windows.com website which has tons and tons of examples: For our app we’ll use this layout. So what will we show? Well, let’s see what sorts of features our app has to offer. I’ll repeat them here: Note planes Add pictures of that plane Notify friends of new spots Share new spots on social media Write down arrival times Write down departure times Write down the runway they take I am sure you can think of some more items but for now we'll use these. In the hub we’ll show something that represents “Spots”, “Friends”, “Social”. Apparently we have an inner list of spotter-friends that are in the app, while we also have to whole world in social. In the layer below we show something else, depending on what the user choose. When they choose “Spots” we’ll display the last spots, last spots by our friends (so we can actually jump from this category to the one next to it) and so on. When they choose a “spot” (or press the + icon in the App bar, which I’ll talk about next time) they go to the lowest and final level that shows details about that spot, including a picture, date and time and the notes belonging to that entry. You’d be amazed at how easy it is to organize your app this way. If you don’t have enough room in these three layers you probably could easily get away with grouping items. Take a look at our hub: we have three completely different things in one place. If you still can’t fit it all in in a logical and consistent way, chances are you are trying to do too much in this app. Go back to your mission statement, determine if it is specific enough and if your feature list helps that statement or makes it unclear. Go ahead. Give it a go! Next time we’ll talk about the look and feel, the charms and the app-bar….

    Read the article

  • JSR 355 Final Release, and moves JCP to version 2.9

    - by heathervc
    JSR 355, JCP EC Merge, passed the JCP EC Final Approval Ballot on 13 August 2012, with 14 Yes votes, 1 abstain (1 member did not vote) on the SE/EE EC, and 12 yes votes (2 members were not eligible to vote) on the ME EC.  JSR 355 posted a Final Release this week, moving the JCP program version to JCP 2.9.  The transition to a merged EC will happen after the 2012 EC Elections, as defined in the Appendix B of the JCP (pasted below), and the EC will operate under the new EC Standing Rules. In the previous version (2.8) of this Process Document there were two separate Executive Committees, one for Java ME and one for Java SE and Java EE combined. The single Executive Committee described in this version of the Process Document will be implemented through the following process: The 2012 annual elections will be held as defined in JCP 2.8, but candidates will be informed that if they are elected their term will be for only a single year, since all candidates must stand for re-election in 2013. Immediately after the 2012 election the two ECs will be merged. Oracle and IBM's second seats will be eliminated, resulting in a single EC with 30 members. All subsequent JSR ballots (even for in-progress JSRs) will then be voted on by the merged EC. For the 2013 annual elections three Ratified and two Elected Seats will be eliminated, thereby reducing the EC to 25 members. All 25 seats will be up for re-election in 2013. Members elected in 2013 will be ranked to determine whether their initial term will be one or two years. The 50% of Ratified and 50% of Elected members who receive the most votes will serve an initial two-year term, while all others will serve an initial one year term. All members elected in 2014 and subsequently will serve a two-year term. For clarity, note that the provisions specified in this version of the Process Document regarding a merged EC will apply to subsequent ballots on all existing JSRs, whether or not the Spec Leads of those JSRs chose to adopt this version of the Process Document in its entirety. <end of Appendix> Also of note:  the materials and minutes from the July EC meeting and the June EC Meeting are now available--following the July EC Meeting, Samsung and SK Telecom lost their EC seats. The June EC meeting also had a public portion--the audio from the public portion of the EC meeting are now posted online.  For Spec Leads there is also the recording of the EG Nominations call.

    Read the article

  • Multiple X screens without Xinerama

    - by Kalle Elmér
    Is there some way to move Windows between separate X screens in Ubuntu 12.04? With 3 monitors on 2 GPUs, using Twinview is not an option. Xinerama allows me to drag windows between all monitors, but it also means that I lose all hardware acceleration. I don't expect to be able to drag windows between the screens and have them occupy two monitors simultaneously, but there must be some way to transfer a program while it is running. At the moment, I don't even know how to select which monitor it will use when launched. Any pointers?

    Read the article

  • How to tune touchpad for smaller area?

    - by M.L.
    I just purchased the Asus Zenbook UX31A and I require some advice regarding tuning the touchpad. Basically my problem is that the mouse buttons are withing the touch area and I would like it to be disabled. Such that only part of the touch area is recognized. I have attached an image to clarify the idea. Basically only the red area in the image should be recognized. I have two reasons for this request. Firstly, I like to hold my thumb on the mouse button instead of holding it in the air. This is problematic, because touchpad thinks that I'm trying some kind of two finger gesture all the time and cursor stops moving. Clicking is also problematic since every finger movement -- regardless of how insignificant it might be -- on the "button" will move the cursor. I have been playing with some synclient settings like AreaBottomEdge etc. They have some effect but then I cannot highlight text or right button stops working. How could I solve this?

    Read the article

< Previous Page | 178 179 180 181 182 183 184 185 186 187 188 189  | Next Page >