Search Results

Search found 20388 results on 816 pages for 'nvidia current'.

Page 34/816 | < Previous Page | 30 31 32 33 34 35 36 37 38 39 40 41  | Next Page >

  • 3 Monitors powered from a Nvidia fx580

    - by user26067
    I have a NVidia FX580 card in a Dell Precision T1500. The outputs of the back of the card are DVI and 2x displayport. I tried to run 3 screens from the DVI + 2x displayport to dvi adaptors, but the card won't have it... It'll only run 2 of these screens at the same time. http://www.nvidia.com/object/product_quadro_fx_580_us.html says; **Display Support:** Dual Link DVI-I 1 DisplayPort 2 # of Digital Outputs 3 ( 2 out of 3 active at a time ) # of Analog Outputs 1 To me this reads that it will be able to power 2xdvi monitors and 1xvga monitor. Anyone care to confirm or speculate?

    Read the article

  • Sharp flat panel scaling with Nvidia drivers

    - by Brecht Machiels
    I have a Samsung 226BW flat panel with a 1680x1050 native resolution. As my PC is rather dated (Athlon XP 2600+ and GeForce 6600 GT), I need to run more recent games (but still old) on a lower resolution. Unfortunately, scaling low resolutions to 1680x1050 results in a very blurry image (bilinear scaling). I have created a custom resolution of 840x525 in the Nvidia control panel. Technically, this resolutions allows perfect upscaling to 1680x1050 without the need for bilinear interpolation. Unfortunately, the Nvidia driver always seems to do bilinear scaling, again resulting in a blurry image. However, I seem to remember that I did obtain crisp images using this resolution in the past (before a Windows re-install). Maybe only some driver versions support integer upscaling without bilinear filtering? Or perhaps there are other solutions?

    Read the article

  • nVidia driver is not installed for x64 Windows 7 (Vaio VGN-SZ460N)

    - by Sergey Osypchuk
    I just upgraded memory to 4GB for Sony Vaio VGN-SZ460N laptop and trying to install Windows 7 x64. All works fine with default drivers, except video - I am not able to set resolution more than 1024x768. Technical specification says that this model is using: Chipset : Intel® 945GM Processor : NVIDIA® GeForce® Go 7400 notebook graphics processing unit (GPU) and Intel® Graphics Media Accelerator 950 I was trying to use Intel 945GM driver, for Windows 7 Ultimate x64 and get error: This computer does not meet the mininum requirements for installing the software Also, I was trying to download nVidia GeForce Go 7400 x64 but it says that there is no compatible hardware. Actually, it has disclaimer that those drivers doesn't work for Sony vaio and recomendation to use OEM website. Sony website has only 32-bit version of driver which says that it cannot be installed and x64 should be used... Any ideas?

    Read the article

  • How can I use 2 monitors plus laptop with my Dell e6420 w/ Nvidia nvs 4200m

    - by KallDrexx
    I have just hooked up a 2nd external monitor to my Dell e6420 laptop with a Nvidia NVS 4200m graphics card, running Windows 8 64 bit. However, the computer won't let me have both monitors and the laptop display active at the same time. I installed the latest Nvidia graphics drivers (310.70) but it claims that my GPU can only support up to 2 monitors. Nivdia's website implies differently (as does various other laptops around the office). The monitors are connected both via DVI to my dell docking station that has multiple DVI ports. Both monitors are working correctly, I just can't get all 3 working together. Attempting to download the driver from dell fails, as their driver installer is broken apparently Any ideas?

    Read the article

  • cuda install in ubuntu13.10?

    - by hexiangpeng
    the cuda_install_.log show ERROR: Unable to build the NVIDIA kernel module. ERROR: Installation has failed. Please see the file '/var/log/nvidia-installer.log' for details. You may find suggestions on fixing installation problems in the README available on the Linux driver download page at www.nvidia.com. The driver installation is unable to locate the kernel source. Please make sure that the kernel source packages are installed and set up correctly. and the other .log show ^ /tmp/selfgz3964/NVIDIA-Linux-x86-319.37/kernel/nv-i2c.c: In function ‘nv_i2c_del_adapter’: /tmp/selfgz3964/NVIDIA-Linux-x86-319.37/kernel/nv-i2c.c:327:14: error: void value not ignored as it ought to be osstatus = i2c_del_adapter(pI2cAdapter); ^ make[3]: * [/tmp/selfgz3964/NVIDIA-Linux-x86-319.37/kernel/nv-i2c.o] ?? 1 make[2]: * [module/tmp/selfgz3964/NVIDIA-Linux-x86-319.37/kernel] ?? 2 NVIDIA: left KBUILD. nvidia.ko failed to build! make[1]: * [module] ?? 1 make: * [module] ?? 2 - Error. ERROR: Unable to build the NVIDIA kernel module. ERROR: Installation has failed. Please see the file '/var/log/nvidia-installer.log' for details. You may find suggestions on fixing installation problems in the README available on the Linux driver download page at www.nvidia.com. i don't understand

    Read the article

  • Current Technologies

    - by Charles Cline
    I currently work at the University of Kansas (KU) and before that Stanford University, to be particular the Stanford Linear Accelerator Center (SLAC).  Collaborating with various Higher Ed institutions the past several years has shown a marked increase in the Microsoft side of the house.  To give you an idea of our current environment, here are some of the things we (Enterprise Systems) have been working on the past two years I’ve been at KU: Migrated from Novell to Active Directory (AD), although we’re still leveraging Novell for IDM.  We currently have 550,000+ objects in AD, and we still have several departments to bring in. Upgraded from Exchange 2003 to Exchange 2010 and Forefront Online Protection for Exchange (FOPE) Implemented SCCM 2007 for Windows systems management Implemented central file storage using EMC products for the backend, using CIFS as the frontend Restructuring AD domains and Forests to decrease the administrative overhead and provide a primary authentication mechanism for the entire University Determining Key Performance Indicators for AD and Exchange Implemented SCOM 2007 to monitor AD and Exchange Implemented Confluence for collaboration within IT and other technology providers at the University Implemented Data Protection Manager (DPM) for backup of AD and Exchange Built a test and QA environment to better facilitate upcoming changes to the environment Almost ready to raise the AD domain level to 2008 R2   I’m sure I’m missing things, and my next post will be some of the things we’re getting ready for – like Centrify to provide AD for OS X and Linux systems.  If anyone would like more info on a particular area, please drop me a line.  I’d be happy to discuss.

    Read the article

  • OpenGL: Attempt to allocate a texture to big for the current hardware

    - by AnonymousMan
    I'm getting the following error: java.io.IOException: Attempt to allocate a texture to big for the current hardware at org.newdawn.slick.opengl.InternalTextureLoader.getTexture(InternalTextureLoader.java:320) at org.newdawn.slick.opengl.InternalTextureLoader.getTexture(InternalTextureLoader.java:254) at org.newdawn.slick.opengl.InternalTextureLoader.getTexture(InternalTextureLoader.java:200) at org.newdawn.slick.opengl.TextureLoader.getTexture(TextureLoader.java:64) at org.newdawn.slick.opengl.TextureLoader.getTexture(TextureLoader.java:24) The image I'm trying to use is 128x128. System.out.println(GL11.glGetInteger(GL11.GL_MAX_TEXTURE_SIZE)); I get: 32. 32??!! My graphics card is AMD Radeon HD 7970M with 2048 MB GDDR5 RAM, I can run all the latest games in 1080p and 60fps with no problem, and those textures sure as hell doesn't look like they are 32x32 pixels to me! How can I fix this? -- Edit: Here's the chaos code I use to init OpenGL: Display.setDisplayMode(new DisplayMode(500,500)); Display.create(); if (!GLContext.getCapabilities().OpenGL11) { throw new Exception("OpenGL 1.1 not supported."); } Display.setTitle("Game"); glMatrixMode(GL_PROJECTION); glLoadIdentity(); GLU.gluPerspective(45, 1, 0.1f, 5000); Mouse.setGrabbed(true); glMatrixMode(GL_MODELVIEW); glLoadIdentity(); glEnable(GL_TEXTURE_2D); glClearColor(0, 0, 0, 0); glEnable(GL_DEPTH_TEST); glDepthFunc(GL_LEQUAL); glHint(GL_PERSPECTIVE_CORRECTION_HINT, GL_NICEST); glBlendFunc(GL_SRC_ALPHA, GL_ONE_MINUS_SRC_ALPHA); glEnable(GL_BLEND); glEnable(GL_POINT_SMOOTH); glEnable(GL_LINE_SMOOTH); glEnable(GL_POLYGON_SMOOTH); glEnable(GL_POLYGON_OFFSET_FILL); glShadeModel(GL_SMOOTH); Display is a LWJGL thing, it makes the OpenGL context and the window. Anyway, I don't think there's anything in the init code that can help me but you never know...

    Read the article

  • Highlight current page tab [migrated]

    - by Jose David Garcia Llanos
    I am making a website, currently i am setting the highlight tab for current page, a particular page is not highlighting its tab and i have checked the code about 5 times but i cant find anything wrong with it. the website is auto-sal.es Here is the code: style.css body#home a.hometab, body#cars a.cartab, body#feedback a.feedtab, body#contact a.contacttab, body#members a.memberstab {background: #7D0000;} contactus.html <body id="contact"> navigation <ul id="menu"> <li><a href="index.html" target="_self" class="hometab">Home</a></li> <li><a href="cars.html" target="_self" class="cartab">Cars</a></li> <li><a href="feedback.html" target="_self" class="feedtab">Feedback</a></li> <li><a href="contactus.html" target="_self" class="cotacttab">Contact Us</a></li> <li><a href="members.html" target="_self" class="memberstab">Members</a></li> </ul> Again, the issue is that it is not highlighting the tab for contact us

    Read the article

  • The current state of a MERGE Destination for SSIS

    - by jamiet
    Hugo Tap asked me on Twitter earlier today whether or not there existed a SSIS Dataflow Destination component that enabled one to MERGE data into a table rather than INSERT it. Its a common request so I thought it might be useful to summarise the current state of play as regards a MERGE destination for SSIS. Firstly, there is no MERGE destination component in the box; that is, when you install SSIS no MERGE Destination will be available. That being said the SSIS team have made available a MERGE destination component via Codeplex which you can get from http://sqlsrvintegrationsrv.codeplex.com/releases/view/19048. I have never used it so cannot vouch for its usefulness although judging by some of the reviews you might not want to set your expectations too high. Your mileage may vary.   In the past it has occurred to me that a built-in way to provide MERGE from the SSIS pipeline would be highly valuable. I assume that this would have to be provided by the database into which you were merging hence in March 2010 I submitted the following two requests to Connect: BULK MERGE (111 votes at the time of writing) [SSIS] BULK MERGE Destination (15 votes) If you think these would be useful feel free to vote them up and add a comment. Lastly, this one is nothing to do with SSIS but if you want to perform a minimally logged MERGE using T-SQL Sunil Agarwal has explained how at Minimal logging and MERGE statement. @Jamiet

    Read the article

  • Current SPARC Architectures

    - by Darryl Gove
    Different generations of SPARC processors implement different architectures. The architecture that the compiler targets is controlled implicitly by the -xtarget flag and explicitly by the -arch flag. If an application targets a recent architecture, then the compiler gets to play with all the instructions that the new architecture provides. The downside is that the application won't work on older processors that don't have the new instructions. So for developer's there is a trade-off between performance and portability. The way we have solved this in the compiler is to assume a "generic" architecture, and we've made this the default behaviour of the compiler. The only flag that doesn't make this assumption is -fast which tells the compiler to assume that the build machine is also the deployment machine - so the compiler can use all the instructions that the build machine provides. The -xtarget=generic flag tells the compiler explicitly to use this generic model. We work hard on making generic code work well across all processors. So in most cases this is a very good choice. It is also of interest to know what processors support the various architectures. The following Venn diagram attempts to show this: A textual description is as follows: The T1 and T2 processors, in addition to most other SPARC processors that were shipped in the last 10+ years supported V9b, or sparcvis2. The SPARC64 processors from Fujitsu, used in the M-series machines, added support for the floating point multiply accumulate instruction in the sparcfmaf architecture. Support for this instruction also appeared in the T3 - this is called sparcvis3 Later SPARC64 processors added the integer multiply accumulate instruction, this architecture is sparcima. Finally the T4 includes support for both the integer and floating point multiply accumulate instructions in the sparc4 architecture. So the conclusion should be: Floating point multiply accumulate is supported in both the T-series and M-series machines, so it should be a relatively safe bet to start using it. The T4 is a very good machine to deploy to because it supports all the current instruction sets.

    Read the article

  • Solved: Operation is not valid due to the current state of the object

    - by ChrisD
    We use public static methods decorated with [WebMethod] to support our Ajax Postbacks.   Recently, I received an error from a UI developing stating he was receiving the following error when attempting his post back: {   "Message": "Operation is not valid due to the current state of the object.",   "StackTrace": "   at System.Web.Script.Serialization.ObjectConverter.ConvertDictionaryToObject(IDictionary`2 dictionary, Type type, JavaScriptSerializer serializer, Boolean throwOnError, Object& convertedObject)\r\n   at System.Web.Script.Serialization.ObjectConverter.ConvertObjectToTypeInternal(Object o, Type type, JavaScriptSerializer serializer, Boolean throwOnError, Object& convertedObject)\r\n   at System.Web.Script.Serialization.ObjectConverter.ConvertObjectToTypeMain(Object o, Type type, JavaScriptSerializer serializer, Boolean throwOnError, Object& convertedObject)\r\n   at System.Web.Script.Serialization.JavaScriptObjectDeserializer.DeserializeInternal(Int32 depth)\r\n   at System.Web.Script.Serialization.JavaScriptObjectDeserializer.DeserializeDictionary(Int32 depth)\r\n   at System.Web.Script.Serialization.JavaScriptObjectDeserializer.DeserializeInternal(Int32 depth)\r\n   at System.Web.Script.Serialization.JavaScriptObjectDeserializer.DeserializeDictionary(Int32 depth)\r\n   at System.Web.Script.Serialization.JavaScriptObjectDeserializer.DeserializeInternal(Int32 depth)\r\n   at System.Web.Script.Serialization.JavaScriptObjectDeserializer.BasicDeserialize(String input, Int32 depthLimit, JavaScriptSerializer serializer)\r\n   at System.Web.Script.Serialization.JavaScriptSerializer.Deserialize(JavaScriptSerializer serializer, String input, Type type, Int32 depthLimit)\r\n   at System.Web.Script.Serialization.JavaScriptSerializer.Deserialize[T](String input)\r\n   at System.Web.Script.Services.RestHandler.GetRawParamsFromPostRequest(HttpContext context, JavaScriptSerializer serializer)\r\n   at System.Web.Script.Services.RestHandler.GetRawParams(WebServiceMethodData methodData, HttpContext context)\r\n   at System.Web.Script.Services.RestHandler.ExecuteWebServiceCall(HttpContext context, WebServiceMethodData methodData)",   "ExceptionType": "System.InvalidOperationException" }   Goggling this error brought me little support.  All the results talked about increasing the aspnet:MaxJsonDeserializerMembers value to handle larger payloads.  Since 1) I’m not using the asp.net ajax model and 2) the payload is very small, this clearly was not the cause of my issue. Here’s the payload the UI developer was sending to the endpoint: {   "FundingSource": {     "__type": "XX.YY.Engine.Contract.Funding.EvidenceBasedFundingSource,  XX.YY.Engine.Contract",     "MeansType": 13,     "FundingMethodName": "LegalTender",   },   "AddToProfile": false,   "ProfileNickName": "",   "FundingAmount": 0 } By tweaking the JSON I’ve found the culprit. Apparently the default JSS Serializer used doesn’t like the assembly name in the __type value.  Removing the assembly portion of the type name resolved my issue. { "FundingSource": { "__type": "XX.YY.Engine.Contract.Funding.EvidenceBasedFundingSource", "MeansType": 13, "FundingMethodName": "LegalTender", }, "AddToProfile": false, "ProfileNickName": "", "FundingAmount": 0 }

    Read the article

  • dist-upgrade runs and completes but current kernel stays the same

    - by jaguare22
    I think that my system is staying at an older kernel version. It seems to update when I run dist-upgrade but the current kernel version doesn't change. Is it possible the system is set to install new kernel updates but only load an older version at start up? $ uname -a Linux HTPC 3.2.0-32-generic #51-Ubuntu SMP Wed Sep 26 21:32:50 UTC 2012 i686 athlon i386 GNU/Linux $ dpkg --list | grep linux-image ii linux-image-3.2.0-32-generic 3.2.0-32.51 Linux kernel image for version 3.2.0 on 32 bit x86 SMP ii linux-image-3.2.0-32-generic-pae 3.2.0-32.51 Linux kernel image for version 3.2.0 on 32 bit x86 SMP ii linux-image-3.2.0-32-virtual 3.2.0-32.51 Linux kernel image for version 3.2.0 on 32 bit x86 Virtual Guests dpkg -l 'linux-*' | sed '/^ii/!d;/'"$(uname -r | sed "s/\(.*\)-\([^0-9]\+\)/\1/")"'/d;s/^[^ ]* [^ ]* \([^ ]*\).*/\1/;/[0-9]/!d' linux-headers-3.2.0-36 linux-headers-3.2.0-36-generic-pae linux-headers-3.2.0-37 linux-headers-3.2.0-37-generic-pae linux-headers-3.2.0-38 linux-headers-3.2.0-38-generic-pae linux-headers-3.2.0-39 linux-headers-3.2.0-39-generic-pae linux-headers-3.2.0-40 linux-headers-3.2.0-40-generic-pae linux-headers-3.2.0-41 linux-headers-3.2.0-41-generic-pae linux-headers-3.2.0-43 linux-headers-3.2.0-43-generic-pae linux-headers-3.2.0-44 linux-headers-3.2.0-44-generic-pae linux-headers-3.2.0-45 linux-headers-3.2.0-45-generic-pae linux-headers-3.2.0-48 linux-headers-3.2.0-48-generic-pae

    Read the article

  • Video memory buswidth vs video memory Bandwidth

    - by Mixxiphoid
    My current video card (9600GT) is dying and I'm searching for a new video card. Between acquiring my current one and now, I got a lot more knowledge about hardware and I want to use that to pick my new card. So I decided to not just buy some popular card blindly, but to search for a card able to handle my hardware requirements. I searched the specs at the NVidia site for the GT640 and was confused by the memory section and some questions raised. My current card's memory bus width is 256bit and has 1GB of memory. I checked Google about the importance of bus width. And all the links basically said the same 'The higher the number the more potential simultaneously traffic can be transferred'. This was already clear to me, yet there are currently a lot of new cards which are considered better than my current one with a lower bus width. To go in more detail about my question I copied the memory info from the NVidia site: GT 640 GT640 GDDR5 Memory Specs: Memory Clock 1.8 Gbps 5.0 Gbps Standard Memory Config 2048 MB 1024 MB Memory Interface DDR3 GDDR5 Memory Interface Width 128-bit 64-bit Memory Bandwidth (GB/sec) 28.5 40.0 What puzzled me is that the Memory Bandwidth seems to me the most important part, yet the lower bus width has the higher 'performance'. Is this due to the fact the memory interface is GDDR5 and is therefore able to have a higher memory clock speed (5Gbps)? If I am to buy a new video card, should I check the bus width? Memory clock? Bandwith? Amount of memory? My current card ahs 1GB memory, so I was searching for a 2GB memory card, but now I'm not so sure any more whether that is really 'better'. My main question: To me it seems that memory performance is made up by the combination of bus width and frequency. Is this true? If yes, why are there so many sites telling me I need to get a card with a high bus width? If no, then what IS important when it goes about memory performance on a video card. NOTE: The memory bandwidth is (almost) never displayed on vendor sites. How can I determine which card is better without knowing the bandwith?

    Read the article

  • Getting Current with Visual Studio 2010 for Web Developers

    - by plitwin
    I don't know about you, but I find it kind of crazy at times figuring out if I have the latest of everything there is for the Visual Studio 2010 developer from Microsoft. (This does not include any third-party components, just recommended updates from Microsoft.) And the be honest, the msn.microsoft.com and asp.net sites are not that helpful in figuring this out.In an effort to help, I have enumerated here what the latest VS 2010 setup should include, complete with download links. When you install everything here, you will be able to develop ASP.NET 4.0 Web Forms and ASP.NET MVC 3 applications and web sites in addition to the other stuff your version of Visual Studio supports (e.g., Silverlight, WPF, etc.). These downloads will also include NuGet and the Entity Framework 4.1, so there is no need to download this software separately.Visual Studio 2010. First of all, you need to purchase and install Visual Studio 2010 itself. For the free Express version, you can download it from Visual Web Developer 2010 ExpressVisual Studio Service Pack 1 (released Spring 2011).This is a must-have download that fixes a bunch of bugs and a number of enhancements too including preliminary support for HTML5 and CSS3. See #4 below for better support of these web technologies. Download and install from VS 2010 SP1 download page. You can find details on the features of the service pack here. ASP.NET MVC3 Tools Update (released Spring 2011)If you are using ASP.NET MVC 3, then you should also download install this update for Visual Studio from ASP.NET MVC3 Tools Update download page. This update improves Visual Studio's support for MVC 3, including better scaffolding, NuGet, Entity Framework 4.1, and more. A good overview of the updates can be found in Phil Haack's blog post.Web Standards Update for Microsoft Visual Studio 2010 SP1 (released June 2011)This is an update to VS 2010 SP1 that "brings VS 2010 intellisense & validation as close to W3C specification as we could get via means of an extension". Download and install from Web Standards Update download page. A good description of the changes can be found in the Visual Web Developer Team blog post.Note: I don't control these download pages, so it is possible they will change. If so, I will do my best to update these links. This information was current as of June 24, 2011.

    Read the article

  • Getting current time in milliseconds

    - by user90293423
    How to get the current time in milliseconds? I'm working on a hacking simulation game and when ever someone connects to another computer/NPC, a login screen popups with a button on the side called BruteForce. When BruteForce is clicked, what i want the program to do is, calculate how many seconds cracking the password is going to take based on the player's CPU speed but that's the easy part. The hard part is i want to enter a character in the password's box every X milliseconds based on a TimeToCrack divided by PasswordLength formula. But since i don't know how to find how many milliseconds have elapsed since the second has passed, the program waits until the CurrentTime is higher than the TimeBeforeTheLoopStarted + HowLongItTakesToTypeaCharacter which is always going to be a second. How would you handle my problems? I've commented the game breaking part. std::vector<QString> hardware = user.getHardware(); QString CPU = hardware[0]; unsigned short Speed = 0; if(CPU == "OMG SingleCore 1.8GHZ"){ Speed = 2; } const short passwordLength = password.length(); /* It's equal to 16 */ int Time = passwordLength / Speed; double TypeSpeed = Time / passwordLength; time_t t = time(0); struct tm * now = localtime(&t); unsigned short EndTime = (now->tm_sec + Time) % 60; unsigned short CurrentTime = 0; short i = passwordLength - 1; do{ t = time(0); now = localtime(&t); CurrentTime = now->tm_sec; do{ t = time(0); now = localtime(&t); }while(now->tm_sec < CurrentTime + TypeSpeed); /* Highly flawed */ /* Do this while your integer value is under this double value */ QString tempPass = password; tempPass.chop(i); ui->lineEdit_2->setText(tempPass); i--; }while(CurrentTime != EndTime);

    Read the article

  • How to force a resolution in linux?

    - by hi
    I have a HP LP3065 that requires a dual link dvi cable to go up to 2560x1600 resolution. However, I do not have that cable, but I want to use 1920x1200 resolution. However, my nVidia Display Settings (also system display settings) will only go up to 1280x1024, which looks horribly pixelated on the 30". How do I force the 1920x1200 resolution? I tried adding the mode into my xorg.conf file, but it still would not take it. I know my vid card can do 1920x1200 since it works on the 24" monitor with the same dvi cable. Here are my specs: Fedora 12 Nvidia Quadro NVS 420 Intel Xenon E5530 cpu 6gig memory Thanks

    Read the article

  • Best linux distro for cuda development

    - by user18151
    Can someone suggest the best linux distro for CUDA development. The reason why I'm asking is that I tried installing the latest cuda SDK in Fedora 12, and it was a real pain in the neck. It took me 8 hours to get the nouveau driver removed and the nvidia-driver installed. After that somehow the OS decides to act up and blow up the /var/log/message file to 9 GB and eat up all my remaining space, with strange errors. I don't even understand what happened further, but my Nvidia drives don't work anymore. Please don't flame me, I'm NOT a windows fanboy or anything. I've been using Linux since 2002, and actually like it. Its just my personal experience. Would be really helpful for positive suggestions. Fanboys, please stay aside. Thanks in advance.

    Read the article

  • Ubuntu 9.10 x86_32 with PAE causes graphics issues, mainly slowness

    - by widgisoft
    I've tried the 64bit version and found I was constantly hitting a brick wall trying to get 32bit stuff to run; I'd previously used PAE on 9.04 without any issues so figured I'd give it a shot. However, on 9.10 it seems PAE or the process of enabling PAE breaks the nvidia drivers/module somewhat as performance is terrible; I can't even enable desktop effects and there's lots of artifacting on random controls. Disabling the pae image and rebooting fixes the issue however I'm then stuck with "only" 2.7Gb of ram and unable to use the full 8gb that's installed. Is there somthing special I need to do when using pae and nVidia drivers or should I be using 64bit and just figure out how to force run 32bit packages? :-p

    Read the article

  • 2 GPUs (for multi-monitor setup)

    - by A Dwarf
    I have a machine to which I need to connect a second monitor. This machine currently hosts a ATI Radeon HD 4770 with two DVI adapters. The second monitor to be hooked to this machine is a VGA monitor. Until I find a DVI-VGA adpater, I'm thinking installing a second graphics card which does have a VGA adapter, an NVidia card. Can I successfully run both AMD and NVidia cards on this machine? Any specifics I must be aware of, like how to ensure the ATI card to be the main card? Operating system is Windows 7

    Read the article

  • Rollback driver in Windows via command line

    - by ultrasawblade
    I have a remote system in which I've updated the Nvidia graphics driver, but now RDPDD.dll, which I'm guessing stands for RDP Display Driver, will not load (found this out via Event Viewer). I can use Sysinternal's Psexec to execute commands though. I'm sure something has gone very wrong and the user of this system (a CAD engineer) won't be able to use his system normally. I've tried two remote reboots and that has not resolved the issue. So I would like to use the "Roll Back" option for the Nvidia driver in the device manager. Is there a command-line way of doing this?

    Read the article

  • Can't install CUDA drivers for GeForce GT555M

    - by saeed
    I've just bought a new Asus n55 laptop. It has 2 graphics cards from Intel and NVIDIA. But when I try to install CUDA's developer driver for my GPU I get this error: "This graphics driver could not find compatible graphics hardware". I have downloaded both of the following files but both of them get mentioned error: Developer Drivers for WinVista and Win7 (270.81) Notebook Developer Drivers for WinVista and Win7 (275.33) How can I fix this problem? Actually how can I develop CUDA programs on my NVIDIA GPU?

    Read the article

  • Are these symptoms of a video card is dying?

    - by K Cloud
    What I am getting from my PC: Scrambled text in the boot menu. The scrambled text dissapears and everthing goes normal after several restarts, or the computer has worked for some time. Display goes blank showing, windows kernel mode driver stopped responding. Sometimes the windows just hangs, I need to restart the pc for that. There used to be some scrambled colors in my windows, so I cleaned up my video carda and port and reinstalled the video card, things are something right. The pc runs normally for hours in safe mode, as diffault display driver works at that case. My PC specs: Core 2 Duo processor 2 GB RAM NVIDIA EN210 SILENT GPU Windows 8 Drivers details: NVIDIA drivers - v331.65, told to be fully compatible with Win8 Currently I am installing Windows 7 and will be putting the older version of driver v320 to test, yet I'm really confused weather my GPU is dying or still on go.

    Read the article

  • DisplayPort to DVI not working on Quadro FX 580

    - by kaosvid
    I have a PNY NVIDIA Quadro FX 580 graphics card with 1x DVI and 2x DisplayPorts. The DVI port works fine with both my Viewsonic monitors but I cannot get either of the DPs to work using the supplied DP to DVI adapter; all I get is a "no signal" on either monitor when connected to either DP port. The NVIDIA Control Panel shows that the second monitor is not connected when in fact it is. How do I get the second monitor to work? System: Windows XP Professional 32-bit Asus P5Q motherboard Core 2 Duo E8500 CPU 4GB PC8500 RAM

    Read the article

< Previous Page | 30 31 32 33 34 35 36 37 38 39 40 41  | Next Page >