Search Results

Search found 2101 results on 85 pages for 'nvidia geforce'.

Page 43/85 | < Previous Page | 39 40 41 42 43 44 45 46 47 48 49 50  | Next Page >

  • HDMI & Display Port stopped work on 11.10

    - by dizzy
    After upgraded two laptops to 11.10, HDMI and Display ports stopped to work. Symptoms on each (btw. it used to work with 11.04 on both): laptop Dell Inspiron 1525 (HDMI, Intel GMX 3100): after HDMI cable is plugged in, screen is corrupted (no panel, no icons), system is unresponsive, TV set receives some signal, but only blue screen and some regular ticks can be heard. Unplugging the cable system recovers. No logs were checked. Thinkpad W510 (DisplayPort, NVidia). Simple "Screens" utility does not recognizes TV set, but this is something to do with the differences between Nvidia driver API and the one expected from the utility, as far I could spot on the net. However, using Nvidia-settings, TV is recognized, but cannot be enabled and used. Beside that, touch pad freezes after HDMI2DisplayPort connector is plugged in the laptop (not immediately, but after few seconds - probably after some handshake with the TV set crashes). It is strange that no such bug reports can be found on the net. So, I guess it is something wrong on our laptops only, but would appreciate some hints (i.e. any known changes recently related to HDMI, Display Port, X-Windows, kernel... wherever I should take a look and fix the issue).

    Read the article

  • GPGPU

    WhatGPU obviously stands for Graphics Processing Unit (the silicon powering the display you are using to read this blog post). The extra GP in front of that stands for General Purpose computing.So, altogether GPGPU refers to computing we can perform on GPU for purposes beyond just drawing on the screen. In effect, we can use a GPGPU a bit like we already use a CPU: to perform some calculation (that doesn’t have to have any visual element to it). The attraction is that a GPGPU can be orders of magnitude faster than a CPU.WhyWhen I was at the SuperComputing conference in Portland last November, GPGPUs were all the rage. A quick online search reveals many articles introducing the GPGPU topic. I'll just share 3 here: pcper (ignoring all pages except the first, it is a good consumer perspective), gizmodo (nice take using mostly layman terms) and vizworld (answering the question on "what's the big deal").The GPGPU programming paradigm (from a high level) is simple: in your CPU program you define functions (aka kernels) that take some input, can perform the costly operation and return the output. The kernels are the things that execute on the GPGPU leveraging its power (and hence execute faster than what they could on the CPU) while the host CPU program waits for the results or asynchronously performs other tasks.However, GPGPUs have different characteristics to CPUs which means they are suitable only for certain classes of problem (i.e. data parallel algorithms) and not for others (e.g. algorithms with branching or recursion or other complex flow control). You also pay a high cost for transferring the input data from the CPU to the GPU (and vice versa the results back to the CPU), so the computation itself has to be long enough to justify the overhead transfer costs. If your problem space fits the criteria then you probably want to check out this technology.HowSo where can you get a graphics card to start playing with all this? At the time of writing, the two main vendors ATI (owned by AMD) and NVIDIA are the obvious players in this industry. You can read about GPGPU on this AMD page and also on this NVIDIA page. NVIDIA's website also has a free chapter on the topic from the "GPU Gems" book: A Toolkit for Computation on GPUs.If you followed the links above, then you've already come across some of the choices of programming models that are available today. Essentially, AMD is offering their ATI Stream technology accessible via a language they call Brook+; NVIDIA offers their CUDA platform which is accessible from CUDA C. Choosing either of those locks you into the GPU vendor and hence your code cannot run on systems with cards from the other vendor (e.g. imagine if your CPU code would run on Intel chips but not AMD chips). Having said that, both vendors plan to support a new emerging standard called OpenCL, which theoretically means your kernels can execute on any GPU that supports it. To learn more about all of these there is a website: gpgpu.org. The caveat about that site is that (currently) it completely ignores the Microsoft approach, which I touch on next.On Windows, there is already a cross-GPU-vendor way of programming GPUs and that is the DirectX API. Specifically, on Windows Vista and Windows 7, the DirectX 11 API offers a dedicated subset of the API for GPGPU programming: DirectCompute. You use this API on the CPU side, to set up and execute the kernels that run on the GPU. The kernels are written in a language called HLSL (High Level Shader Language). You can use DirectCompute with HLSL to write a "compute shader", which is the term DirectX uses for what I've been referring to in this post as a "kernel". For a comprehensive collection of links about this (including tutorials, videos and samples) please see my blog post: DirectCompute.Note that there are many efforts to build even higher level languages on top of DirectX that aim to expose GPGPU programming to a wider audience by making it as easy as today's mainstream programming models. I'll mention here just two of those efforts: Accelerator from MSR and Brahma by Ananth. Comments about this post welcome at the original blog.

    Read the article

  • IE8: intense flickering for Flash in Windows 7 RTM

    - by 280Z28
    Whenever I view a page with Flash on it (example www.fox.com), if I move my mouse around the page flickers intensely - like siezure inducing flicker. Is this a known issue, and is there a way to fix it? Windows 7 RTM x64 IE8 Flash Player, Acrobat Reader, Shockwave Player are all I installed Consistent across 3 very different machines (AMD ZM-82 + Radeon HD3200 laptop, Core i7 + NVidia GT220 desktop, P4D + NVidia 6400GT desktop), but all with the above software.

    Read the article

  • Do i need to join Graphics Card with CPU fan

    - by Mirage
    Initially i had inno 3d 256MB Nvidia GTS graphics card. I also had another Big FAN above the processor (Vendor put in quad core) In that card there was one cable which was joined with that FAN. Now i have changed the CARD to 1GB Nvidia GT9600 . But there is no pins to join the fan with that card. Is it ok . i don't know why old card was joined with FAN

    Read the article

  • Cannot exit X server, restart, shutdown or drop to tty when VGA monitor active

    - by terdon
    I have a strange problem. If I connect an external VGA monitor to my laptop, exiting the X environment in any way crashes the computer. For example, say I am working with my two monitors (the laptop's and one connected to my VGA port) active. Hitting Ctrl+Alt+F Key should take me down to a tty. What actually happens is that the VGA screen goes blank, as you would expect, but the laptop screen, although still on, shows nothing. I know the screen is on because it is slightly more illuminated than when it is off. When in this state, I can do nothing to regain access to the machine. I have tried: Ctrl+F Key (and even Ctrl+Alt+F Key, just in case) combinations and none seem to have any effect. Ctrl+Alt+Del : Nothing Magic SysRq key: Nothing Blindly typing my username and password and trying to reboot/shutdown or restart GDM or MDM: Nothing The only thing that works is a hard reset. The exact same behavior occurs when kiling the X server through Ctrl+Alt+Backspace, rebooting or shutdown. There is no difference if I reboot/shutdown/log out using the WM's graphical menu or if I use the shutdown or rebootcommands. It is also not WM-dependent. I have the same problem using Cinnamon, Gnome 3, MATE and xfce4. It is, however, VGA dependent. I have tried connecting another VGA monitor and have the same problem. I do not, however, have this problem if a screen is connected to the DisplayPort. It is, therefore, a VGA specific issue. To make things even stranger, this only occurs when both screens are active. If either the laptop screen or the VGA monitor is inactive the problem goes away. Finally, this problem arose when I installed the latest Linux Mint Debian (LMDE). It did not occur with the previous release of LMDE. I am not sure what has changed since I used the same kernel version in both releases (I had upgraded the kernel while on the previous release) and, I think, the same nvidia drivers. Oh, and yes, I have updated the nvidia driver. Hardware: Dell M4500 laptop CPU: Intel Core i7 RAM: 8GB Graphics: nVidia GT216 [Quadro FX 880M] Software: LMDE, kernel 3.2.0-2-amd64 Xorg: 1.11.4 nVidia kernel: 295.20-1+3.2.9-1 Possibly relevant files: /var/log/Xorg.0.log ~/.xsession-errors Does anyone have any ideas how to fix this? Thanks in advance for any help.

    Read the article

  • What DX level does my graphics card support? Does it go to 11?

    - by Daniel Moth
    Recently I run into a situation that I have run into quite a few times. Someone encounters a machine and the question arises: "Is there a DirectX 11 card in this machine?". Typically the reason you are interested in that is because cards with DirectX 11 drivers fully support DirectCompute (and by extension C++ AMP) for GPGPU programming. The driver specifically is WDDM (1.1 on Windows 7 and Windows 8 introduces WDDM 1.2 with cool new capabilities). There are many ways for figuring out if you have a DirectX11 card, so here are the approaches that you can use, with a bonus right at the end of the post. Run DxDiag WindowsKey + R, type DxDiag and hit Enter. That is the DirectX diagnostic tool, which unfortunately, only tells you on the "System" tab what is the highest version of DirectX installed on your machine. So if it reports DirectX 11, that doesn't mean you have a DX11 driver! The "Display" tab has a promising "DDI version" label, but unfortunately that doesn't seem to be accurate on the machines I've tested it with (or I may be misinterpreting its use). Either way, this tool is not the one you want for this purpose, although it is good for telling you the WDDM version among other things. Use the Microsoft hardware page There is a Microsoft Windows 7 compatibility center, that lists all hardware (tip: use the advanced search) and you could try and locate your device there… good luck. Use Wikipedia or the hardware vendor's website Use the Wikipedia page for the vendor cards, for both nvidia and amd. Often this information will also be in the specifications for the cards on the IHV site, but is is nice that wikipedia has a single page per vendor that you can search etc. There is a column in the tables for API support where you can see the DirectX version. Check if it is one of these recommended DX11 cards You may not have a DirectX 11 card and are interested in purchasing one. While I am in no position to make recommendations, I will list here some cards from two big IHVs that we know are DirectX 11 capable. Some AMD (aka ATI) cards Low end, inexpensive DX11 hardware: Radeon 5450, 5550, 6450, 6570 Mid range (decent perf, single precision): Radeon 5750, 5770, 6770, 6790 High end (capable of double precision): Radeon 5850, 5870, 6950, 6970 Single precision APUs: AMD E-Series APUs AMD A-Series APUs Some NVIDIA cards Low end, inexpensive DX11 hardware: GeForce GT430, GT 440, GT520, GTS 450 Quadro 400, 600 Mid-range (decent perf, single precision): GeForce GTX 460, GTX 550 Ti, GTX 560, GTX 560 Ti Quadro 2000 High end (capable of double precision): GeForce GTX 480, GTX 570, GTX 580, GTX 590, GTX 595 Quadro 4000, 5000, 6000 Tesla C2050, C2070, C2075 Get the DirectX SDK and run DirectX Caps Viewer Download and install the June 2010 DirectX SDK. As part of that you now have the DirectX Capabilities Viewer utility (find it in your start menu by searching for "DirectX Caps Viewer", the filename is DXCapsViewer.exe). It will list all your devices (emulated, and real hardware ones) under the first node. Expand the hardware entries and then expand again the Direct3D 11 folder. If you see D3D_FEATURE_LEVEL_11_ under that, then your card supports feature level 11 which means it supports DirectCompute and C++ AMP. In the following screenshot of one of my old laptops, the card only goes to feature level 10. Run a utility from the web that just tells you! Of course, writing some C++ AMP code that enumerates accelerators and lists the ones that are capable is trivial. However that requires that you have redistributed the runtime, so a more broadly applicable approach is to use the DX APIs directly to enumerate the DX11 capable cards. That is exactly what the development lead for C++ AMP has done and he describes and shares that utility at this post. Comments about this post by Daniel Moth welcome at the original blog.

    Read the article

  • What DX level does my graphics card support? Does it go to 11?

    - by Daniel Moth
    Recently I run into a situation that I have run into quite a few times. Someone encounters a machine and the question arises: "Is there a DirectX 11 card in this machine?". Typically the reason you are interested in that is because cards with DirectX 11 drivers fully support DirectCompute (and by extension C++ AMP) for GPGPU programming. The driver specifically is WDDM (1.1 on Windows 7 and Windows 8 introduces WDDM 1.2 with cool new capabilities). There are many ways for figuring out if you have a DirectX11 card, so here are the approaches that you can use, with a bonus right at the end of the post. Run DxDiag WindowsKey + R, type DxDiag and hit Enter. That is the DirectX diagnostic tool, which unfortunately, only tells you on the "System" tab what is the highest version of DirectX installed on your machine. So if it reports DirectX 11, that doesn't mean you have a DX11 driver! The "Display" tab has a promising "DDI version" label, but unfortunately that doesn't seem to be accurate on the machines I've tested it with (or I may be misinterpreting its use). Either way, this tool is not the one you want for this purpose, although it is good for telling you the WDDM version among other things. Use the Microsoft hardware page There is a Microsoft Windows 7 compatibility center, that lists all hardware (tip: use the advanced search) and you could try and locate your device there… good luck. Use Wikipedia or the hardware vendor's website Use the Wikipedia page for the vendor cards, for both nvidia and amd. Often this information will also be in the specifications for the cards on the IHV site, but is is nice that wikipedia has a single page per vendor that you can search etc. There is a column in the tables for API support where you can see the DirectX version. Check if it is one of these recommended DX11 cards You may not have a DirectX 11 card and are interested in purchasing one. While I am in no position to make recommendations, I will list here some cards from two big IHVs that we know are DirectX 11 capable. Some AMD (aka ATI) cards Low end, inexpensive DX11 hardware: Radeon 5450, 5550, 6450, 6570 Mid range (decent perf, single precision): Radeon 5750, 5770, 6770, 6790 High end (capable of double precision): Radeon 5850, 5870, 6950, 6970 Single precision APUs: AMD E-Series APUs AMD A-Series APUs Some NVIDIA cards Low end, inexpensive DX11 hardware: GeForce GT430, GT 440, GT520, GTS 450 Quadro 400, 600 Mid-range (decent perf, single precision): GeForce GTX 460, GTX 550 Ti, GTX 560, GTX 560 Ti Quadro 2000 High end (capable of double precision): GeForce GTX 480, GTX 570, GTX 580, GTX 590, GTX 595 Quadro 4000, 5000, 6000 Tesla C2050, C2070, C2075 Get the DirectX SDK and run DirectX Caps Viewer Download and install the June 2010 DirectX SDK. As part of that you now have the DirectX Capabilities Viewer utility (find it in your start menu by searching for "DirectX Caps Viewer", the filename is DXCapsViewer.exe). It will list all your devices (emulated, and real hardware ones) under the first node. Expand the hardware entries and then expand again the Direct3D 11 folder. If you see D3D_FEATURE_LEVEL_11_ under that, then your card supports feature level 11 which means it supports DirectCompute and C++ AMP. In the following screenshot of one of my old laptops, the card only goes to feature level 10. Run a utility from the web that just tells you! Of course, writing some C++ AMP code that enumerates accelerators and lists the ones that are capable is trivial. However that requires that you have redistributed the runtime, so a more broadly applicable approach is to use the DX APIs directly to enumerate the DX11 capable cards. That is exactly what the development lead for C++ AMP has done and he describes and shares that utility at this post. Comments about this post by Daniel Moth welcome at the original blog.

    Read the article

  • Intel GMA 4500MHD flickering in opengl

    - by Aaron
    I have an application that uses OpenGL for its display. When this application is run on a laptop with an Intel GMA 4500MHD on Windows XP the OpenGL content appears to flicker/shudder between 2 frames. I have experienced the exact same problem when running this application on nVidia Quadro FX hardware, however with nVidia the problem can be corrected by disabling the "Unified Back Buffer" (UBB) feature in the driver. Does anybody know how I can disable the corresponding feature in the Intel driver?

    Read the article

  • Which is a better graphics card

    - by michael
    Hi, Can someone please give me advice regarding which of the following is a better graphics cards? Radeon HD4650 1GB NVIDIA Geforce GT240M 1GB Or which brand is better in general? Radeon? or Nvidia? Thank you.

    Read the article

  • Ubuntu 10.04 says wired network disconnected, but it works

    - by mistily
    I've installed ubuntu 10.04, hoping to get rid of the 9.10 errors. (Nvidia driver didn't work quite properly, had some problems with clicking in flash, and I had -and still have- wired network working, but ubuntu said that it's disconnected) So the Nvidia driver is better now, but the network error is the same. I still have a working wired network connection, but when it starts says that it is disconnected. Notification area tells me the same thing. Why is that?

    Read the article

  • Intel GMA 4500MHD flickering in opengl

    - by Aaron
    I have an application that uses OpenGL for its display. When this application is run on a laptop with an Intel GMA 4500MHD on Windows XP the OpenGL content appears to flicker/shudder between 2 frames. I have experienced the exact same problem when running this application on nVidia Quadro FX hardware, however with nVidia the problem can be corrected by disabling the "Unified Back Buffer" (UBB) feature in the driver. Does anybody know how I can disable the corresponding feature in the Intel driver?

    Read the article

  • Windows 7 losing one of my displays after restart

    - by j_kubik
    I have an Intel DZ68BC motherboard with Intel HD graphics card using two monitors (on DVI and on HDMI » VGA). My friend asked me to test if his NVIDIA graphics card works well on my computer (at his it was doing some trouble), so I inserted it in my computer, installed the NVIDIA driver and it worked quite well. Then I removed it, uninstalled everything NVIDIA-related I could find and switched monitors back to my Intel card. Since then after every system start/restart, the system sees only monitor on HDMI » VGA connector, completely ignoring the DVI monitor. I noticed that installing the Intel video drivers causes the system to recognize the second monitor if I don't immediately reboot. After a reboot, the system recognizes only the HDMI » VGA monitor. I also tried starting in safe-mode and using DriveSweeper to remove the remains of NVIDIA drivers. While it seems that some drivers were removed, the situation didn't change. Now I am out of ideas and I really wouldn't like to reinstall the system (again...). I also tried restoring the system to the state before this whole story, but it also didn't change anything. EDIT: I am still trying to troubleshoot this problem. The only point that I could start was driver re-instalation. I traced down the part that restores right settings to a call: C:\Users\Jarek\Desktop\GFX_Win7_64_8.15.10.2696\x64\Drv64.exe -driverinf "C:\Users\Jarek\Desktop\GFX_Win7_64_8.15.10.2696\Graphics\igdlh64.inf" -flags 20 -keypath "Software\Intel\Difx64" This call fixes my displays, and as workaround, I will add it for now to my autorun. I am still looking for better solution anyway... EDIT2: Using DriverView i made a list of currently used drivers both before and after fixing my display using above command. Then i compared logs: No drivers were removed by fixing command. Drivers added by fixing command: MS Remote Access serial network driver (asyncmac.sys) security processor (spsys.sys) Drivers that changed base address (indicates driver-reload?) Canonical Display Driver (cdd.dll) Intel Graphics Kernel Mode Driver (igdkmd64.sys) Monitor Driver (monitor.sys) Added drivers seem rather unrelated to the problem to me, reloaded drivers are just a cnsequence of installing new driver file so there is not much to go here... I really cannot make heads or tails out of it...

    Read the article

  • GPU Temperature comparisons between similar GPUs of the same line

    - by White Phoenix
    As a general rule of thumb, if there are two similar GPUs from the same line (i.e. an NVIDIA GTX 260 vs a NVIDIA GTX 280), will the less powerful or more powerful GPU run hotter? This is assuming all other factors stay the same - hardware configuration, cooling setup, ingame settings, etc. Or does this depend entirely upon the GPU itself? I do remember reading that the GTX 280 in this example was terribly inefficient with power and cooling.

    Read the article

  • My OpenCL kernel is slower on faster hardware.. But why?

    - by matdumsa
    Hi folks, As I was finishing coding my project for a multicore programming class I came up upon something really weird I wanted to discuss with you. We were asked to create any program that would show significant improvement in being programmed for a multi-core platform. I’ve decided to try and code something on the GPU to try out OpenCL. I’ve chosen the matrix convolution problem since I’m quite familiar with it (I’ve parallelized it before with open_mpi with great speedup for large images). So here it is, I select a large GIF file (2.5 MB) [2816X2112] and I run the sequential version (original code) and I get an average of 15.3 seconds. I then run the new OpenCL version I just wrote on my MBP integrated GeForce 9400M and I get timings of 1.26s in average.. So far so good, it’s a speedup of 12X!! But now I go in my energy saver panel to turn on the “Graphic Performance Mode” That mode turns off the GeForce 9400M and turns on the Geforce 9600M GT my system has. Apple says this card is twice as fast as the integrated one. Guess what, my timing using the kick-ass graphic card are 3.2 seconds in average… My 9600M GT seems to be more than two times slower than the 9400M.. For those of you that are OpenCL inclined, I copy all data to remote buffers before starting, so the actual computation doesn’t require roundtrip to main ram. Also, I let OpenCL determine the optimal local-worksize as I’ve read they’ve done a pretty good implementation at figuring that parameter out.. Anyone has a clue? edit: full source code with makefiles here http://www.mathieusavard.info/convolution.zip cd gimage make cd ../clconvolute make put a large input.gif in clconvolute and run it to see results

    Read the article

  • Dual monitors through 1 HDMI port

    - by Carlos
    I currently have a Dell Studio XPS 13 laptop connected to a 24" HP monitor (w2448hc). Im thinking on getting a second one, however am wondering what i need for the setup (hardware wise). Also I was wondering it there is any down side to it, or something i should be aware of. For example, image quality loss, GPU overloading, or anything important I should know. More than anything Im interested in your advice. Also the monitors do have built in speakers (HDMI sound output), is the sound going to be reproduced by only one monitor or both? Specs Model: Dell Studio XPS 13 OS: Genuine Windows® 7 Home Premium 64-Bit CPU: Intel® CoreTM 2 Duo P8600 (2.4GHz, 3MB L2 Cache, 1067MHz FSB) Chipset: NVIDIA® GeForce® MCP79MX RAM: 4GB 1067MHz DDR3 SDRAM Graphics: SLi NVIDIA® GeForce® 9500M - 256MB Thanks for your advice, if there is anything additional i need to buy an you have a personal preference pass the brand name so i can check it out. Thanks!

    Read the article

  • Dual monitors though 1 HDMI port

    - by Carlos
    I currently have a Dell Studio XPS 13 laptop connected to a 24" HP monitor (w2448hc). Im thinking on getting a second one, however am wondering what i need for the setup (hardware wise). Also I was wondering it there is any down side to it, or something i should be aware of. For example, image quality loss, GPU overloading, or anything important I should know. More than anything Im interested in your advice. Also the monitors do have built in speakers (HDMI sound output), is the sound going to be reproduced by only one monitor or both? Specs Model: Dell Studio XPS 13 OS: Genuine Windows® 7 Home Premium 64-Bit CPU: Intel® CoreTM 2 Duo P8600 (2.4GHz, 3MB L2 Cache, 1067MHz FSB) Chipset: NVIDIA® GeForce® MCP79MX RAM: 4GB 1067MHz DDR3 SDRAM Graphics: SLi NVIDIA® GeForce® 9500M - 256MB Thanks for your advice, if there is anything additional i need to buy an you have a personal preference pass the brand name so i can check it out. Thanks!

    Read the article

  • 12.04 Black screen with blinking cursor

    - by Junaid
    I have read this post My computer boots to a black screen, what options do I have to fix it? UBUNTU 12.04 works well with livecd but fails to start after complete installation, black screen with a blinking cursor. I tried holding shift key and edit grub options but holding shift did not do anything. I give up, any suggestion. I have built in intel graphics card and an nvidia card in PCI. I have tried by removing nvidia card as well. All I see is a black screen with a blinking cursor. My system is Dell optiplex GX260 with 1GB ram and 2.4 GHz P4 processor

    Read the article

  • Acer Revo (ION platform) + Maverick + 5.1 surround over HDMI

    - by Oli
    I've had a turbulent relationship with my media centre box. Every upgrade I perform on it seems to bring a brand new set of audio issues (the opposite of my desktop where things seem to get better and better). It's a Acer Revo 3600. That's basically an low-end Intel Atom chip with a Nvidia 9400M onboard. On paper that's perfect for something like a media centre. But having just upgraded to Maverick, the sound properties box only wants to offer me stereo sound over HDMI. The exact setup goes: Revo - Onkyo AV receiver - LG TV. The Onkyo box strips off the audio (supporting 7.1 -- though we're only using 6 speakers) and feeds the video onto the TV. I'd like to get to a point where Ubuntu thinks it's doing 5.1 over HDMI, upmixing stereo to 6ch and supporting DTS/AC3 (through Boxee). I've had this working before but it's frankly been a bit of a hacktastrophe. The audio chip is recognised as Nvidia MCP79/7A HDMI in alsamixer if that helps.

    Read the article

  • I cannot install a proporietary fglrx driver

    - by Chris
    OK my install failed while installing the amd proprietary driver. I keep seeing references to a Nvidia driver, but I do not have or even want anything pertaining to Nvidia. Here are my latest entries for the jockey.log: 2012-10-11 20:31:24,103 DEBUG: fglrx_updates is not the alternative in use 2012-10-11 20:31:24,119 DEBUG: fglrx.enabled(fglrx_updates): target_alt None current_alt /usr/lib/fglrx/ld.so.conf other target alt None other current alt /usr/lib/fglrx/alt_ld.so.conf 2012-10-11 20:31:24,119 DEBUG: fglrx_updates is not the alternative in use 2012-10-11 20:31:24,322 DEBUG: fglrx.enabled(fglrx): target_alt /usr/lib/fglrx/ld.so.conf current_alt /usr/lib/fglrx/ld.so.conf other target alt /usr/lib/fglrx/alt_ld.so.conf other current alt /usr/lib/fglrx/alt_ld.so.conf 2012-10-11 20:31:24,323 DEBUG: XorgDriverHandler(%s, %s).enabled(): No X.org driver set, not checking 2012-10-11 20:31:24,450 DEBUG: fglrx.enabled(fglrx_updates): target_alt None current_alt /usr/lib/fglrx/ld.so.conf other target alt None other current alt /usr/lib/fglrx/alt_ld.so.conf 2012-10-11 20:31:24,450 DEBUG: fglrx_updates is not the alternative in use 2012-10-11 20:31:24,465 DEBUG: fglrx.enabled(fglrx_updates): target_alt None current_alt /usr/lib/fglrx/ld.so.conf other target alt None other current alt /usr/lib/fglrx/alt_ld.so.conf 2012-10-11 20:31:24,465 DEBUG: fglrx_updates is not the alternative in use

    Read the article

  • How do I get Unity working again after installing the wrong video driver?

    - by Jesse
    First off I did something stupid. I downloaded a Nvidia driver even though I have an integrated chipset. After installation my unity was still working. However, when I restarted my computer I got an error message saying that I can't run unity. I uninstalled the Nvidia driver. I restarted my computer. Unity still does not work. In the terminal I type "unity" and everything looks okay until I get three error messages that say this: Xlib: extension "GLX" missing on display ":0.0". followed by: Compiz (opengl) - Fatal: glXCreateContext failed Compiz (bailer) - Info: Ensuring a shell for your session jesse@jesse-PC:~$ Cannot register the panel shell: there is already one running.

    Read the article

  • La loi de Moore a-t-elle encore un avenir ? Oui, répond un ingénieur, si les développeurs se mettent

    La loi de Moore est elle encore pertinente ? Plus pour les CPU, répond un responsable de Nvidia qui appelle au développement de la programmation parallèle Bill Dally est un des ingénieurs les plus importants de Nvidia. Dans une tribune publiée dans le magasine Forbes, l'ingénieur en chef doute fortement que la loi de Moore puisse encore s'appliquer aux processeurs (CPU) : « Les performances des CPUs ne peuvent plus doubler tous les 18 mois », constate-t-il, « et cela va poser un grave problème à de nombreuses industries qui reposent sur cette croissance des performances ». Optimiste, Bill Dally en tire cependant une raison d'espérer « La bonne nouvelle, c'e...

    Read the article

  • Omni-directional shadow mapping

    - by gridzbi
    What is a good/the best way to fill a cube map with depth values that are going to give me the least amount of trouble with floating point imprecision? To get up and running I'm just writing the raw depth to the buffer, as you can imagine it's pretty terrible - I need to to improve it, but I'm not sure how. A few tutorials on directional lights divide the depth by W and store the Z/W value in the cube map - How would I perform the depth comparison in my shadow mapping step? The nvidia article here http://http.developer.nvidia.com/GPUGems/gpugems_ch12.html appears to do something completely different and use the dot of the light vector, presumably to counter the depth precision worsening over distance? He also scales the geometry so that it fits into the range -.5 +.5 - The article looks a bit dated, though - is this technique still reasonable? Shader code http://pastebin.com/kNBzX4xU Screenshot http://imgur.com/54wFI

    Read the article

  • Looking for literature about graphics pipeline optimization

    - by zacharmarz
    I am looking for some books, articles or tutorials about graphics architecture and graphics pipeline optimizations. It shouldn't be too old (2008 or newer) - the newer, the better. I have found something in [Optimising the Graphics Pipeline, NVIDIA, Koji Ashida] - too old, [Real-time rendering, Akenine Moller], [OpenGL Bindless Extensions, NVIDIA, Jeff Bolz], [Efficient multifragment effects on graphics processing units, Louis Frederic Bavoil] and some internet discussions. But there is not too much information and I want to read more. It should contain something about application, driver, memory and shader units communication and data transfers. About vertices and attributes. Also pre and post T&L cache (if they still exist in nowadays architectures) etc. I don't need anything about textures, frame buffers and rasterization. It can also be about OpenGL (not about DirecX) and optimizing extensions (not old extensions like VBOs, but newer like vertex_buffer_unified_memory).

    Read the article

  • A question about "add-apt-repository"(gcc 4.7 or bumblebee)

    - by girlkoo
    I install ubuntu 12.04 in my notebook and I use c++. I want to use c++11,so I want to upgrade my gcc to gcc4.7. I do like this: sudo add-apt-repository ppa:ubuntu-toolchain-r/test sudo apt-get update sudo apt-get install gcc-4.7 sudo apt-get install g++-4.7 but when I run "sudo apt-get install gcc-4.7", it just can find gcc-4.7-base not gcc-4.7, and when I run "sudo apt-get install g++-4.7", it can find nothing... Today, I want to install the "bumblebee" and do like this: sudo add-apt-repository ppa:bumblebee/stable sudo apt-get update sudo apt-get install bumblebee bumblebee-nvidia sudo reboot but when I run "sudo apt-get install bumblebee bumblebee-nvidia", it tells me the packages can't be find. The ways I methoded are specially use to ubuntu 12.04, the links is: Installing gcc 4.7 https://wiki.ubuntu.com/Bumblebee#Installation What can I do? Thank you!My email is [email protected]

    Read the article

< Previous Page | 39 40 41 42 43 44 45 46 47 48 49 50  | Next Page >