Search Results

Search found 6563 results on 263 pages for 'graphics tablet'.

Page 48/263 | < Previous Page | 44 45 46 47 48 49 50 51 52 53 54 55  | Next Page >

  • OpenGL and switchable graphic cards

    - by Orcun
    I use a laptop and this laptop has readon AMD Radeon HD 6470M and onboad graphic card. When I run fglrxinfo, I get this error: X Error of failed request: BadRequest (invalid request code or no such operation) Major opcode of failed request: 136 (GLX) Minor opcode of failed request: 19 (X_GLXQueryServerString) Serial number of failed request: 12 Current serial number in output stream: 12 Is it a problem ? Because of I reason I can't use opengl. Because, I can't run any opengl applications.

    Read the article

  • Enable Compiz on Alienware M11x R2

    - by user3307
    I want to enable compiz on my Alienware m11x r2 but when I try to I get a message that I have to enable Nvidia propietary drivers and I cant enable them because it will cause problems with next startup because the nvidia is the discrate GPU of my laptop, I had search a little on this here and here. But I seem to not get it working right, if someone can help me I will thank you alot! I have also tried the vga-swticheroo thing with no results at all...

    Read the article

  • ATI card - cannot set refresh rate higher than 60 Hz

    - by KubaV
    I am trying to set refresh rate to 85 Hz or 75 Hz at 1024x768 and 1280x1024. On Windows this works, so my card and monitor have 100% support. xrandr does not work - it gives me configure crtc 1 failed I use FGLRX. Please help Monitor:CRT - HP 92 (manual http://bizsupport2.austin.hp.com/bc/docs/support/SupportManual/lpv09818/lpv09818.pdf) G. card: AMD Radeon HD 6450 Sorry for bad english, I am Czech.

    Read the article

  • How do I get a new license for gDEBugger after the 1 free year?

    - by Byte56
    I downloaded the gDEBugger from gremedy over a year ago, with their one year free license. The license has since expired and their site says that I'll be presented with the option for 1 year free license the first time I run it after install. This doesn't happen when re-installing, it just tells me the license has expired. How do I get a new license? I use this regularly for debugging shader problems and performance testing my game.

    Read the article

  • How To Be Your Own Personal Clone Army (With a Little Photoshop)

    - by Eric Z Goodnight
    Maybe you’ve always wanted more of yourself. Or maybe you’ve always thought you could be your own best friend! Regardless of your reasons, here’s how to duplicate yourself with some clever photograph tricks and either Photoshop or GIMP. How To Be Your Own Personal Clone Army (With a Little Photoshop) How To Properly Scan a Photograph (And Get An Even Better Image) The HTG Guide to Hiding Your Data in a TrueCrypt Hidden Volume

    Read the article

  • How do I convert matrices intended for OpenGL to be compatible for DirectX?

    - by gardian06
    I have finished working through the book "Game Physics Engine Development 2nd Ed" by Millington, and have got it working, but I want to adapt it to work with DirectX. I understand that D3D9+ has the option to use either left handed, or right handed convention, but I am unsure about how to return my matrices to be usable by D3D. The source code gives returning OpenGL column major matrices (the transpose of the working transform matrix shown below), but DirectX is row major. For those unfamiliar for the organization of the matrices used in the book: [r11 r12 r13 t1] [r21 r22 r23 t2] [r31 r32 r33 t3] [ 0 0 0 1] r## meaning the value of that element in the rotation matrix, and t# meaning the translation value. So the question in short is: How do I convert the matrix above to be easily usable by D3D? All of the documentation that I have found simply states that D3D is row major, but not where to put what elements so that it is usable by D3D in terms of rotation, and translation elements.

    Read the article

  • Doesn't boot after installing fglrx for ati 5650

    - by elivanja
    Hi, I'm a complete noob here, I recently bought an HP 3161 with an ATI 5650 hd card, when I installed 10.04 LTS everything was working fine until I was prompted to activate proprietary driver fglrx, after I did this I couldn't reboot (not even in recovery mode) until I reinstalled 10.04, what can I do? I would appreciate any help since I haven't been able to find anyone with the same problem, thanks.

    Read the article

  • How to install Radeon Open Source Driver in Ubuntu 12.04

    - by Clintonio
    I have an MSI Radeon HD 7850 card. MSI provide no Linux support, and the AMD driver gives me an Unsupported Hardware watermark that does not show up in screenshots. I uninstalled the proprietary driver and found that Unity 3D no longer works, and I really only picked Ubuntu because of the excellent shortcuts provided in the recent versions. Trying to locate the open source driver has been a pain for me. According to the documentation I have read it is pre-included in Ubuntu 12.04, yet there is no Unity 3D support. I tried using the newest drivers from AMD's website, but that also had the watermark. Does anyone know how I can get the open source driver? If not, does anyone know how to remove the watermark. I'm on the verge of going back to Windows.

    Read the article

  • hpdv6 6080 two graphic card

    - by Taher
    My laptop has two graphic card one is ati radeon 6770m and other intel sandybridge. i install ati driver from repository but i can't select ati as my graphic card after installing ati ubuntu lose 3d mode how can i set graphic card? and switch between two graphic cards? or how can i set one of them? my laptop is hp dv6 6080 and i add blacklist radeon to /etc/modprobe.d/blacklist.conf and add below lines to /etc/rc.local file: modprobe radeon echo OFF /sys/kernel/debug/vgaswitcheroo/switch exit 0

    Read the article

  • Installing Ubuntu, Switches of to intergrated Graphic Card

    - by Clenn
    I got a problem. I'm willing to dual boot ubuntu 12.04 on my desktop. After i made a bootable USB ubuntu on it i'm going to install it. I boot it and I choose Install ubuntu roughly said. The problem is the following, It seems that when i press enter to start the install he switches of to my intergrated graphic card. I looked up in my bios what my primary graphic card selection is and it was on PCI Express (the option it supossed to be). So I'm clueless right now. I don't know how to fix this because I'm quite learning linux atm (school and by myself). Your faithfully, Clenn

    Read the article

  • Crash on shutdown when using second monitor

    - by plua
    I have recently added a second monitor to my setup. Instead of one VGA monitor, I now have one VGA and one DVI monitor hooked up to the computer. The displays work just fine. However, ever since this monitor was added Ubuntu (10.10) does not shut down properly. I go to shutdown and it starts shutting down but just hangs and I need to manually shut it down. This has been like this for weeks and I feel not comfortable shutting it down this way over and over again. Has anybody any idea what this can be? Or, also useful, does anybody know how to get some info on why this crash occurs? Any logs that can be used for this? Thanks! Paul

    Read the article

  • My laptop overheat on ubuntu 13.04

    - by Sina
    Overheating maybe is exaggerating word, but it's temperature is more in comparison with windows. My Graphic Card has failed for 2 time and I've paid money for each time. The problem makes me disappointed of keeping on ubuntu. Whenever I want to install "nvidia-current" my resolution becomes too low in screen and sometimes it just shows a dark screen. Is there any other Linux distribution which be more compatible with Hardware, specially Graphic Card? My Laptop is Dell Vostro 1510. And once This problem happened in 12.04 and another time days ago :(

    Read the article

  • How can I create an orthographic display that handles different screen dimensions?

    - by Piku
    I'm trying to create an iPad/iPhone game using GLES2.0 that contains a 3D scene with a heads-up-display/GUI overlaid on the top. However, this problem would also apply if I were to port my game to a computer and run the game in a resizable window, or allow the user to change screen resolutions... When trying to make the 2D GUI/HUD work I've made the assumption that all I'm really doing is drawing a load of 2D textured 'quads' on the screen and am trying to treat the orthographic projection as an old-style 2D display with 0,0 in the upper left and screenWidth,ScreenHeight in the lower right. This causes me all sorts of confusion when I rotate my ipad into Landscape mode since I can't work out what to put into my projection and modelview matrices to turn everything around the right way. It also gets messy if I want to support the iPad's large screen, an iPhone or a Retina display since I have to then draw three sets of textures for everything and work out which ones to use. Should I be trying to map the 2D OpenGL co-ords 1:1 with the screen? While typing out this question it occurs to me that I could keep my origin in the centre, still running -1/+1 along the axes. This would let me scale my 2D content appropriately on the different screen sizes, but wouldn't I end up with the textures being scaled and possibly losing quality? I'm using OpenGLES 2.0 and have a matrix library that has equivalents to the GLES1.1 glOrthof() and glFrustrum() calls.

    Read the article

  • Working with vectors and transformations

    - by user29163
    I am going to write an graphical 2D application that allows user to create polygons and transform them through transformation such as rotation an so on. I was hoping someone can give pro and cons arguments for the different choices I got in my mind. (Its all in Java btw!) a). Represent vectors by filling matrices with 'real' numbers. This means making a matrix datas tructure that supports multiplication, transposing etc b). Make a own vector class, such that I can make a matrix class that support those vectors.

    Read the article

  • How would I go about implementing a globe-like "ballish" map?

    - by rFactor
    I am new to 3D development and I have this idea of having the game world like our globe is - a ball. So, there would be no corners in the map and the game is top-down RTS game. I would like the camera to go on and on and never stop even though you are moving the camera to the same direction all the time. I am not sure if this is even possible, but I would really like to build a globe-like map without borders. Can this be done, and how exactly? I am using XNA, C#, and DirectX. Any tutorials or links or help is greatly appreciated!

    Read the article

  • How can I get SLI working with 295.40?

    - by Steve
    I've been doing a lot of googling these last few hours and I'm not having much luck. Perhaps I don't know exactly what I am looking for. I just recently installed Ubuntu 12.04LTS x86_64. Looks beautiful! I have two GTX470's in SLI, and I am finally migrating my desktop over given the hopeful gaming support as of late. My laptop has been enjoying multiple distros of Ubuntu for a couple years now. However, new problems come with unexplored territory, here. At first, I only had one working monitor of my two. Over on nvidia-xconfig I fixed that, but the only solution that actually worked was twinview. Just recently I read here that twinview is not compatible with SLI. Sweet. When I try to tell it, oh hey, use a separate XScreen, configure it the way I want it, click save to configuration file, enter my password, then a sudo restart lightdm, it's broken. One screen blacks or whites out (Couldn't tell you the specific conditions for each, I'm dubious at this point,) and I get this huge error dialogue box upon login. Something about incompatible resolutions if I remember right. Though I am sure I set the resolutions for each screen correctly. Anyway, when I try to enable SLI (sudo nvidia-xconfig --sli=On) despite the fact it hates twinview, unity breaks. The sidebar is there, but only one screen works, the mouse is trapped running along the left edge of it, and the background of the sidebar is a solid blue. Anyway, this ended up being entirely too verbose, I'm sorry, but could anyone part some wisdom please? It would be appreciated!

    Read the article

  • How do I improve terrain rendering batch counts using DirectX?

    - by gamer747
    We have determined that our terrain rendering system needs some work to minimize the number of batches being transferred to the GPU in order to improve performance. I'm looking for suggestions on how best to improve what we're trying to accomplish. We logically split our terrain mesh into smaller grid cells which are 32x32 world units. Each cell has meta data that dictates the four 256x256 textures that are used for spatting along with the alpha blend data, shadow, and light mappings. Each cell contains 81 vertices in a 9x9 grid. Presently, we examine each cell and determine the four textures that are being used to spat the cell. We combine that geometry with any other cell that perhaps uses the same four textures regardless of spat order. If the spat order for a cell differs, the blend map is adjusted so that the spat order is maintained the same as other like cells and blending happens in the right order too. But even with this batching approach, it isn't uncommon when looking out across an area of open terrain to have between 1200-1700 batch count depending upon how frequently textures differ or have different texture blends are between cells. We are only doing frustum culling presently. So using texture spatting, are there other alternatives that can reduce the batch count and allow rendering to be extremely performance-friendly even under DirectX9c? We considered using texture atlases since we're targeting DirectX 9c & older OpenGL platforms but trying to repeat textures using atlases and shaders result in seam artifacts which we haven't been able to eliminate with the exception of disabling mipmapping. Disabling mipmapping results in poor quality textures from a distance. How have others batched together terrain geometry such that one could spat terrain using various textures, minimizing batch count and texture state switches so that rendering performance isn't negatively impacted?

    Read the article

  • What would be a good opensource software for graphing circles, circular regions, and regions bounded by curves and/or arcs?

    - by Michael Dykes
    Hullo all. I have just started s job with Chegg, and my 1st assignment has me writign solutions for Stewart's Essential Calculus. I am dealing with the chapter on multiple integration, and need a good open-source software that I can easily use to draw regions (domains) that would require multiple integrals: i.e. circular regions, portions of circles, regions bounded by curves and/or arcs, and at some point 3D pictures. In most of these cases, I am not working with an exact equation or perhaps need to draw the region bounded by r (radius) between 1 and 2, and the angle theta bounded between pi/4 and 3 pi/4. I am not too terribly familiar with programs like Corel Draw, but the documents I have from this company suggest Corel Draw. So I think I am looking for an open-source free program like Corel Draw or something similar. Any additional suggestions would also be appreciated. I know I can do most, if not all of this using TikZ, but the learning curve is a bit steep, and at the moment I an on a time constraint. Thanks.

    Read the article

  • How to render Minecraft on the GPU?

    - by l0b0
    Hardware: Intel i7 AMD Radeon HD 6970 SSD with plenty of space 6 GB RAM Software OpenJDK 6, 7, and Oracle Java 7 (reproducible with all three) AMD Catalyst 12.8 and open source driver (reproducible with both) Ubuntu 12.04 x86_64 and older Minecraft 1.3.2 vanilla and older On this setup I am getting rubbish frame rates after a short while of playing, dropping from about 45-55 to 15 in a couple of minutes. CPU use is 40-45 even when rendering the opening screen at 1920x1280, and gameRenderer is using about 90% CPU when playing. Rather than trying to eke out a few more FPS out of an obviously broken rendering pipeline, I really hope to find a solution to make the GPU render Minecraft.

    Read the article

  • After 10.10 -> 11.04 upgrade, can only login via Classic (No Effects)

    - by Ryan P.
    Yesterday I upgraded from 10.10 to 11.04, everything seemed to go okay until immediately after login: the desktop goes into a "corrupted" looking state (similar to having too high resolution set). I can see some kind of movement by moving the mouse around/right clicking, and can enter text terminals via ctrl + alt + f1 It does this in both plain "Ubuntu" and "Ubuntu Classic", and only seems to login/startup properly with Ubuntu Classic (No Effects). I have checked my video card (Radeon X600) and run the unity support test which passes with all "yes" results (Unity supported: yes): /usr/lib/nux/unity_support_test -p I have tried re-installing my Ubuntu desktop: rm -rf .gnome .gnome2 .gconf .gconfd .metacity sudo apt-get remove ubuntu-desktop sudo apt-get install ubuntu-desktop With no success. I can workaround for now with Classic (No Effects), but I'd really like to find the root problem. Any suggestions on what else to try would be appreciated!

    Read the article

  • OGRE 3D: How to create very basic gameworld [on hold]

    - by skiwi
    I'm considering trying around to create an FPS (First person shooter), using the Ogre 3D engine. I have done the Basic Tutorials (except CEGUI), and have read through the Intermediate Tutorial, I understand some of the more advanced concepts, but I'm stuck with very simple concepts. First of all: I would want to use some tiles (square ones, with relative little height) as the floor, I guess I need to set up a loop to get those tiles done. But how would I go about creating those tiles exactly? Like making it to be their own mesh, and then I would need to find some texture. Secondly: I guess I can derive the camera and movement functions from the basic tutorial. But I'll be needing a "soldier" (anything does for now), what is the best way to create a moderately decent looking soldier? (Or obtain a decent one from an open library?) And thirdly: How can I ensure that the soldier is actually walking on the ground, instead of mid air? Will raycasting into the ground + adjust position based on that, suffice?

    Read the article

  • Book suggestions for GUI related topics [closed]

    - by asrijaal
    In the past, I've developed more on the server side, so no big GUI/graphic programming topics have crossed my way so far. I'm switching a little bit to the mobile plattforms (iOS, android) and would like to gather more info on the visual/graphic side. Using the SDK controls/widgets isn't that challenging and I would love to get deeper into the view/presentation layer. Animations, drawing itself, probably OpenGL. Any books that would take me a little deeper in understanding how drawing actually works would be helpful.

    Read the article

  • Calculating a circle or sphere along a vector

    - by Sparky
    Updated this post and the one at Math SE (http://math.stackexchange.com/questions/127866/calculating-a-circle-or-sphere-along-a-vector), hope this makes more sense. I previously posted a question (about half an hour ago) involving computations along line segments, but the question and discussion were really off track and not what I was trying to get at. I am trying to work with an FPS engine I am attempting to build in Java. The problem I am encountering is with hitboxing. I am trying to calculate whether or not a "shot" is valid. I am working with several approaches and any insight would be helpful. I am not a native speaker of English nor skilled in Math so please bear with me. Player position is at P0 = (x0,y0,z0), Enemy is at P1 = (x1,y1,z1). I can of course compute the distance between them easily. The target needs a "hitbox" object, which is basically a square/rectangle/mesh either in front of, in, or behind them. Here are the solutions I am considering: I have ruled this out...doesn't seem practical. [Place a "hitbox" a small distance in front of the target. Then I would be able to find the distance between the player and the hitbox, and the hitbox and the target. It is my understanding that you can compute a circle with this information, and I could simply consider any shot within that circle a "hit". However this seems not to be an optimal solution, because it requires you to perform a lot of calculations and is not fully accurate.] Input, please! Place the hitbox "in" the player. This seems like the better solution. In this case what I need is a way to calculate a circle along the vector, at whatever position I wish (in this case, the distance between the two objects). Then I can pick some radius that encompasses the whole player, and count anything within this area a "hit". I am open to your suggestions. I'm trying to do this on paper and have no familiarity with game engines. If any software folk out there think I'm doing this the hard way, I'm open to help! Also - Anyone with JOGL/LWJGL experience, please chime in. Is this making sense?

    Read the article

  • Is using a dedicated thread just for sending gpu commands a good idea?

    - by tigrou
    The most basic game loop is like this : while(1) { update(); draw(); swapbuffers(); } This is very simple but have a problem : some drawing commands can be blocking and cpu will wait while he could do other things (like processing next update() call). Another possible solution i have in mind would be to use two threads : one for updating and preparing commands to be sent to gpu, and one for sending these commands to the gpu : //first thread while(1) { update(); render(); // use gamestate to generate all needed triangles and commands for gpu // put them in a buffer, no command is send to gpu // two buffers will be used, see below pulse(); //signal the other thread data is ready } //second thread while(1) { wait(); // wait for second thread for data to come send_data_togpu(); // send prepared commands from buffer to graphic card swapbuffers(); } also : two buffers would be used, so one buffer could be filled with gpu commands while the other would be processed by gpu. Do you thing such a solution would be effective ? What would be advantages and disadvantages of such a solution (especially against a simpler solution (eg : single threaded with triple buffering enabled) ?

    Read the article

  • How can I support scrolling when using batched rendering for my tiles?

    - by dardanel
    I have tiled map 100*75 and tiles are 32*32 pixel.I want to use batching for performance .I don't figure it out , because of my game needs scrolling and every frame i draw 22*16 tiles (my screen is 20*16 tile) .I thought that batching tiles for every frame .Is it good or any suggestion? edit :to more clarify I want to use occlusion culling and batching at the same time.I thought that drawing only visible areas and batching them together .But there is a something i couldn't figure out .When scrolling screen with translate matrix , if one row become invisible , I bind new row and batch them again.Every batched objects needs to buffer again.So I batch tiles and buffer to VBO every time when one row become invisible .I don't know these way is efficient or not .This is my question .And i am open to any suggestions.

    Read the article

< Previous Page | 44 45 46 47 48 49 50 51 52 53 54 55  | Next Page >