Search Results

Search found 2379 results on 96 pages for 'nvidia prime'.

Page 9/96 | < Previous Page | 5 6 7 8 9 10 11 12 13 14 15 16  | Next Page >

  • Nvidia Proprietary Drivers

    - by Ben
    I have an X79 motherboard and a GTX 670. I have tried to install every combination of Ubuntu 11.10, 12.04, 12.10, Linux Mint 13 and the Nvidia Driver 304.43, 295.xx, etc. I have read everything and pretty much tried everything...drivers from the site, PPAs, some random scripts, editing the xorg.conf file. It has been four days of futile effort. Should I just wait...or something? Has anyone gotten hardware like this to work?

    Read the article

  • Gnome 3 - Multiple Video Cards - Xinerama -- Forced Fallback Mode

    - by Alvin
    Just installed a 2nd nvidia video card -- previously had gnome 3 working perfectly with 2 monitors on a a single video card using twinview tried a number of things thus far twinview on 1 card + xinerama no xinerama no twinview various manual xorg.conf hacks based on random forums (couple references below) xinerama no twinview with and without Extensions Composite The last one is what I'm using now -- it results in a forced fallback mode with Composite Disable set at the end of xorg.conf via nvidia-settings Section "Extensions" Option "Composite" "Disable" EndSection when I disabled that last snippet it boots to gnome 3 full with the left monitor on a black screen and the middle monitor as primary but non-responsive switching to console mode Ctrl+Alt+F1 and then switching back I get 3 black screens with a mouse that can move around but nothing to interact with issue seems related to OpenGL and the multiple video cards -- I can boot into Unity without issue though my Glx-Dock shows up with the black background as barely shows in the screenshot below indicating the OpenGL is not initiated has anyone had any luck with getting Xinerama to work with Multiple NVidia Video Cards with OpenGL support? Found this in the logs while looking a bit further [ 23.208] (II) NVIDIA(1): Setting mode "nvidia-auto-select+0+0" [ 23.254] (WW) NVIDIA(1): The GPU driving screen 1 is incompatible with the rest of the [ 23.254] (WW) NVIDIA(1): GPUs composing the desktop. OpenGL rendering will be [ 23.254] (WW) NVIDIA(1): disabled on screen 1. [ 23.277] (==) NVIDIA(1): Disabling shared memory pixmaps [ 23.277] (==) NVIDIA(1): Backing store disabled [ 23.277] (==) NVIDIA(1): Silken mouse enabled [ 23.277] (==) NVIDIA(1): DPMS enabled According to this page at the NVidia User Docs http://us.download.nvidia.com/XFree86/Linux-x86/173.14.09/README/chapter-14.html I may be out of luck =( Starting this question with the hopes that others may be able to help debug and perhaps gain answers over time as I really want to get the full gnome 3 back.

    Read the article

  • Checking if an int is prime more efficiently

    - by SipSop
    I recently was part of a small java programming competition at my school. My partner and I have just finished our first pure oop class and most of the questions were out of our league so we settled on this one (and I am paraphrasing somewhat): "given an input integer n return the next int that is prime and its reverse is also prime for example if n = 18 your program should print 31" because 31 and 13 are both prime. Your .class file would then have a test case of all the possible numbers from 1-2,000,000,000 passed to it and it had to return the correct answer within 10 seconds to be considered valid. We found a solution but with larger test cases it would take longer than 10 seconds. I am fairly certain there is a way to move the range of looping from n,..2,000,000,000 down as the likely hood of needing to loop that far when n is a low number is small, but either way we broke the loop when a number is prime under both conditions is found. At first we were looping from 2,..n no matter how large it was then i remembered the rule about only looping to the square root of n. Any suggestions on how to make my program more efficient? I have had no classes dealing with complexity analysis of algorithms. Here is our attempt. public class P3 { public static void main(String[] args){ long loop = 2000000000; long n = Integer.parseInt(args[0]); for(long i = n; i<loop; i++) { String s = i +""; String r = ""; for(int j = s.length()-1; j>=0; j--) r = r + s.charAt(j); if(prime(i) && prime(Long.parseLong(r))) { System.out.println(i); break; } } System.out.println("#"); } public static boolean prime(long p){ for(int i = 2; i<(int)Math.sqrt(p); i++) { if(p%i==0) return false; } return true; } } ps sorry if i did the formatting for code wrong this is my first time posting here. Also the output had to have a '#' after each line thats what the line after the loop is about Thanks for any help you guys offer!!!

    Read the article

  • Unable to use OpenGL or install nVidia driver on openSUSE 12.2

    - by djechelon
    I have an ASUS N76VZ laptop with 12.2 openSUSE and GeForce GT650M card. I found that KDE doesn't allow me to use OpenGL rendering. I tried to install nVidia's driver from script but once it writes the xorg.conf file I'm unable to boot desktop. I have the following errors in system log Oct 30 08:28:13 RAYNOR kdm[2727]: X server died during startup Oct 30 08:28:13 RAYNOR kdm[2727]: X server for display :0 cannot be started, session disabled I noticed that the /etc/X11/xorg.conf backup file was empty, so I renamed the new xorg.conf and left none: the desktop booted!!! How can I fix OpenGL rendering with or without driver installation? [Update]: Xorg.0.log says [ 1434.207] compiled for 4.0.2, module version = 1.0.0 [ 1434.207] Module class: X.Org Server Extension [ 1434.207] (II) NVIDIA GLX Module 304.60 Sun Oct 14 20:44:54 PDT 2012 [ 1434.207] (II) Loading extension GLX [ 1434.207] (II) LoadModule: "record" [ 1434.207] (II) Loading /usr/lib64/xorg/modules/extensions/librecord.so [ 1434.207] (II) Module record: vendor="X.Org Foundation" [ 1434.207] compiled for 1.12.3, module version = 1.13.0 [ 1434.207] Module class: X.Org Server Extension [ 1434.207] ABI class: X.Org Server Extension, version 6.0 [ 1434.207] (II) Loading extension RECORD [ 1434.207] (II) LoadModule: "dri" [ 1434.207] (II) Loading /usr/lib64/xorg/modules/extensions/libdri.so [ 1434.207] (II) Module dri: vendor="X.Org Foundation" [ 1434.207] compiled for 1.12.3, module version = 1.0.0 [ 1434.207] ABI class: X.Org Server Extension, version 6.0 [ 1434.207] (II) Loading extension XFree86-DRI [ 1434.207] (II) LoadModule: "nvidia" [ 1434.208] (II) Loading /usr/lib64/xorg/modules/drivers/nvidia_drv.so [ 1434.208] (II) Module nvidia: vendor="NVIDIA Corporation" [ 1434.208] compiled for 4.0.2, module version = 1.0.0 [ 1434.208] Module class: X.Org Video Driver [ 1434.208] (II) NVIDIA dlloader X Driver 304.60 Sun Oct 14 20:24:42 PDT 2012 [ 1434.208] (II) NVIDIA Unified Driver for all Supported NVIDIA GPUs [ 1434.208] (++) using VT number 8 [ 1434.320] (EE) No devices detected. [ 1434.320] Fatal server error: [ 1434.320] no screens found [ 1434.320] Please consult the The X.Org Foundation support at http://wiki.x.org for help.

    Read the article

  • Switching from Onboard intel to Nvidia Dedicated GPU

    - by Anarkie
    How can I switch from Intel onboard grpahics to Nvidia Dedicated GPU? When I go to windows screen resolution I see intel. I cant change it. I go to Device Manager, I see both Adapters are there and Nvidia is known.I disabled intel, I didnt see any option to set one as primary so I disabled intel, black screen!Reboot and re-enable intel. I right click on the desktop, choose "Nvidia Control Panel" and on 3D options I chose the desired game I want to play, High performance Nvidia, but it didnt switch when I started the game. Then I made preferred GPU in the global settings High performance Nvidia for everything it still didnt change.I understand to save the battery etc. there is a switch option between these two but I dont see this switch when it is necessary, I cant also switch manually?Is there a manual switch FN key?I looked but couldnt find. Why I want to do this? 1) Better game peformance. 2) I want to play an old game from 2002(Diablo 2 LOD), when I start the game there are black bars on the sides, so screen becomes just smaller which I dislike!I heard this is intel's specification to center the display.But instead I would like to scale or expand it to fit widescreen(fullscreen).Which should be possible with Nvidia. My Notebook Specs: Fujitsu Lifebook AH531, Win7 , 64 bit, i5, intel HD graphics onboard, Nvidia GT 525. I didnt install Nvidia later, it was always installed and ready from the moment I turned on the computer first time. How I determined that the cards werent switched when I am playing the game: with the windows key I exited from the game, then looked at screen resolutions menu, still saw intel, also the game was still with black bars.I know intel GPU should enough for Diablo 2 but I am interested in this answer for further games, I dont always play Diablo, what if I install an up to date game for example?Then Intel will not be sufficient.I would like to learn the switch option.

    Read the article

  • ERROR running Bumblebee on 13.10

    - by paul
    I'm trying to get Bumblebee working again after an upgrade to Saucy. Running software with Optirun gives the following output: optirun nvidia-settings [ 45.697126] [ERROR]Cannot access secondary GPU - error: [XORG] (EE) Failed to load /usr/lib/xorg/modules/libglamoregl.so: /usr/lib/xorg/modules/libglamoregl.so: undefined symbol: _glapi_tls_Context [ 45.697179] [ERROR]Aborting because fallback start is disabled. Does anyone know how to fix this? Thanks! :)

    Read the article

  • System reverts to 87Hz refresh rate at every startup after I have installed nvidia drivers

    - by Mohammad Kamil Nadeem
    Every time the system starts the screen's refresh rate reverts to 87Hz which results in a pixelated and flickery screen which I have to manually correct every time by either selecting 60Hz as my refresh rate. I have tried "save to X configuration files" and even tried by making the changes as Root but to no avail as it again reverts to 87Hz on every system startup The Open Source Drivers are Okay for regular Unity but many games don't work on it hence I had to install the nvidia drivers. I have been facing this since the Beta Phase although this is on a fresh installation of 12.04 final release. I am also providing my Xorg.conf file just in case it might help http://paste.ubuntu.com/952196/ Also for some reason Displays shows my CRT monitor as Laptop but on open source drivers it was mentioning it as a 14" CRT only This bug is also present on Edubuntu 12.04 This is not present on Xubuntu 12.04 I had selected to install updates and 3rd party software on the install and was greeted with a correct refresh rate screen on the Boot Up. I like Xubuntu.

    Read the article

  • Often "running in low-graphics mode” with NVidia GeForce GT 430

    - by user42228
    Often (every fifth time perhaps) when I turn on my system, I'm greeted with the "running in low-graphics mode” window. Sometimes all I have to do is reboot and the system will start correctly, sometimes there can be a longer series of failed starts. If I go to the terminal and log in, I can run startx and my desktop will appear without problems. At the end of /var/log/Xorg.1.log I see this: [ 77.459] (EE) Screen(s) found, but none have a usable configuration. [ 77.459] Fatal server error: [ 77.459] no screens found [ 77.459] Please consult the The X.Org Foundation support at http://wiki.x.org Is the NVidia creating the appropriate XOrg configuration dynamically on every boot, or what could explain that the configuration error above only occurs sometimes?

    Read the article

  • Basic shadow mapping fails on NVIDIA card?

    - by James
    Recently I switched from an AMD Radeon HD 6870 card to an (MSI) NVIDIA GTX 670 for performance reasons. I found however that my implementation of shadow mapping in all my applications failed. In a very simple shadow POC project the problem appears to be that the scene being drawn never results in a draw to the depth map and as a result the entire depth map is just infinity, 1.0 (Reading directly from the depth component after draw (glReadPixels) shows every pixel is infinity (1.0), replacing the depth comparison in the shader with a comparison of the depth from the shadow map with 1.0 shadows the entire scene, and writing random values to the depth map and then not calling glClear(GL_DEPTH_BUFFER_BIT) results in a random noisy pattern on the scene elements - from which we can infer that the uploading of the depth texture and comparison within the shader are functioning perfectly.) Since the problem appears almost certainly to be in the depth render, this is the code for that: const int s_res = 1024; GLuint shadowMap_tex; GLuint shadowMap_prog; GLint sm_attr_coord3d; GLint sm_uniform_mvp; GLuint fbo_handle; GLuint renderBuffer; bool isMappingShad = false; //The scene consists of a plane with box above it GLfloat scene[] = { -10.0, 0.0, -10.0, 0.5, 0.0, 10.0, 0.0, -10.0, 1.0, 0.0, 10.0, 0.0, 10.0, 1.0, 0.5, -10.0, 0.0, -10.0, 0.5, 0.0, -10.0, 0.0, 10.0, 0.5, 0.5, 10.0, 0.0, 10.0, 1.0, 0.5, ... }; //Initialize the stuff used by the shadow map generator int initShadowMap() { //Initialize the shadowMap shader program if (create_program("shadow.v.glsl", "shadow.f.glsl", shadowMap_prog) != 1) return -1; const char* attribute_name = "coord3d"; sm_attr_coord3d = glGetAttribLocation(shadowMap_prog, attribute_name); if (sm_attr_coord3d == -1) { fprintf(stderr, "Could not bind attribute %s\n", attribute_name); return 0; } const char* uniform_name = "mvp"; sm_uniform_mvp = glGetUniformLocation(shadowMap_prog, uniform_name); if (sm_uniform_mvp == -1) { fprintf(stderr, "Could not bind uniform %s\n", uniform_name); return 0; } //Create a framebuffer glGenFramebuffers(1, &fbo_handle); glBindFramebuffer(GL_FRAMEBUFFER, fbo_handle); //Create render buffer glGenRenderbuffers(1, &renderBuffer); glBindRenderbuffer(GL_RENDERBUFFER, renderBuffer); //Setup the shadow texture glGenTextures(1, &shadowMap_tex); glBindTexture(GL_TEXTURE_2D, shadowMap_tex); glTexImage2D(GL_TEXTURE_2D, 0, GL_DEPTH_COMPONENT, s_res, s_res, 0, GL_DEPTH_COMPONENT, GL_FLOAT, NULL); glTexParameteri(GL_TEXTURE_2D, GL_TEXTURE_MAG_FILTER, GL_NEAREST); glTexParameteri(GL_TEXTURE_2D, GL_TEXTURE_MIN_FILTER, GL_NEAREST); glTexParameteri(GL_TEXTURE_2D, GL_TEXTURE_WRAP_S, GL_CLAMP_TO_EDGE); glTexParameteri(GL_TEXTURE_2D, GL_TEXTURE_WRAP_T, GL_CLAMP_TO_EDGE); return 0; } //Delete stuff void dnitShadowMap() { //Delete everything glDeleteFramebuffers(1, &fbo_handle); glDeleteRenderbuffers(1, &renderBuffer); glDeleteTextures(1, &shadowMap_tex); glDeleteProgram(shadowMap_prog); } int loadSMap() { //Bind MVP stuff glm::mat4 view = glm::lookAt(glm::vec3(10.0, 10.0, 5.0), glm::vec3(0.0, 0.0, 0.0), glm::vec3(0.0, 1.0, 0.0)); glm::mat4 projection = glm::ortho<float>(-10,10,-8,8,-10,40); glm::mat4 mvp = projection * view; glm::mat4 biasMatrix( 0.5, 0.0, 0.0, 0.0, 0.0, 0.5, 0.0, 0.0, 0.0, 0.0, 0.5, 0.0, 0.5, 0.5, 0.5, 1.0 ); glm::mat4 lsMVP = biasMatrix * mvp; //Upload light source matrix to the main shader programs glUniformMatrix4fv(uniform_ls_mvp, 1, GL_FALSE, glm::value_ptr(lsMVP)); glUseProgram(shadowMap_prog); glUniformMatrix4fv(sm_uniform_mvp, 1, GL_FALSE, glm::value_ptr(mvp)); //Draw to the framebuffer (with depth buffer only draw) glBindFramebuffer(GL_FRAMEBUFFER, fbo_handle); glBindRenderbuffer(GL_RENDERBUFFER, renderBuffer); glBindTexture(GL_TEXTURE_2D, shadowMap_tex); glFramebufferTexture2D(GL_FRAMEBUFFER, GL_DEPTH_ATTACHMENT, GL_TEXTURE_2D, shadowMap_tex, 0); glDrawBuffer(GL_NONE); glReadBuffer(GL_NONE); GLenum result = glCheckFramebufferStatus(GL_FRAMEBUFFER); if (GL_FRAMEBUFFER_COMPLETE != result) { printf("ERROR: Framebuffer is not complete.\n"); return -1; } //Draw shadow scene printf("Creating shadow buffers..\n"); int ticks = SDL_GetTicks(); glClear(GL_DEPTH_BUFFER_BIT); //Wipe the depth buffer glViewport(0, 0, s_res, s_res); isMappingShad = true; //DRAW glEnableVertexAttribArray(sm_attr_coord3d); glVertexAttribPointer(sm_attr_coord3d, 3, GL_FLOAT, GL_FALSE, 5*4, scene); glDrawArrays(GL_TRIANGLES, 0, 14*3); glDisableVertexAttribArray(sm_attr_coord3d); isMappingShad = false; glBindFramebuffer(GL_FRAMEBUFFER, 0); printf("Render Sbuf in %dms (GLerr: %d)\n", SDL_GetTicks() - ticks, glGetError()); return 0; } This is the full code for the POC shadow mapping project (C++) (Requires SDL 1.2, SDL-image 1.2, GLEW (1.5) and GLM development headers.) initShadowMap is called, followed by loadSMap, the scene is drawn from the camera POV and then dnitShadowMap is called. I followed this tutorial originally (Along with another more comprehensive tutorial which has disappeared as this guy re-configured his site but used to be here (404).) I've ensured that the scene is visible (as can be seen within the full project) to the light source (which uses an orthogonal projection matrix.) Shader utilities function fine in non-shadow-mapped projects. I should also note that at no point is the GL error state set. What am I doing wrong here and why did this not cause problems on my AMD card? (System: Ubuntu 12.04, Linux 3.2.0-49-generic, 64 bit, with the nvidia-experimental-310 driver package. All other games are functioning fine so it's most likely not a card/driver issue.)

    Read the article

  • Nvidia Drivers, Ubuntu 12.10, GeForce GT 540m

    - by Stacks
    How do I install Nvidia drivers for Ubuntu 12.10? I have a GeForce GT 540m and have been trying to do this for the past 2 months with no luck. I tried the official repository, then I added the x swat ppa. I've installed Linux-headers-generic and Linux-source, and virtually every solution google has to offer. Throughout the course of this I must have done at least 25 fresh installs of Ubuntu; does anybody else(with a GT 540m) have this problem?

    Read the article

  • Linux kernel 3.5.0-18 update is not working with NVIDIA driver 304.60

    - by Janos Fejos
    I have Ubuntu 12.10 64 bit and a Nvidia card GT430 and installed 304.60 Linux driver manually and it worked beautifully until the latest Ubuntu update which got modified to 3.5.0-18 Kernel. Everything has gone: launcher, windows I can't move them it got USELESS. I tried to run the previous kermel 3.5.0-17 with which the driver was working well. I gto the same desktop. USeless - So the problem is something else I guess but it has to do with the update that's for sure- Any idea what to do? Now I have to switch back to Windows 7 which I don't want in order to able to work until an update will come up.

    Read the article

  • gnome-shell failed to launch when dual monitor detected, nvidia card

    - by Terry Hu
    I installed a clean precise. I like gnome-shell so I installed it. It worked well with single monitor. But It does not work well when I go to office, plug my external monitor onto my laptop. I am using nvidia driver. Another colleague is using the same laptop, he can work well with dual monitors. His laptop is ATI graphic card. When I plug my extended monitor cable, I choose GNOME, but actually it goes to gnome 2 classic desktop environment, not GNOME 3. When I disconnect the extended monitor, restart X, I can login with GNOME3. Do I have to choose between GNOME3 and extended monitor? It's a hard choose.

    Read the article

  • ubuntu 12.10 - nvidia gtx 550 ti proprietary drivers issue

    - by Phill
    I have installed Ubuntu 12.1 and own the nvidia gtx 550 ti (no integrated video). Installed it just fine but the buttons and text seem to get scrambled if using the nouveau drivers I guess. I tried installing the proprietary drivers manualy and from the Additional Drivers tab in software sources but every time I do that after it restarts the menus and icons dissapear an the only way getting them back is to revert to the nouveau drivers. I've tried noumerous attempts ending up reinstalling it. Any help would be apreciated, thanks.

    Read the article

  • Ubuntu does not work with Nvidia drivers, gives random squares after logging in

    - by jimbo
    I tried installing both with wubi and by booting from a usb drive. After I tried both it was the same. Ubuntu tried to login, but after a while, a weird pattern with squarely shapes appeared. At one point, after using recovery mode and then the low graphics one (i don't remember how it was called exactly) i managed to get in. It still didn't display at correct resolution, and the mouse was invisible at times. However, after rebooting once, because i tried downloading the nvidia drivers, i didn't manage to get in again, so i have the same problem... I have no idea what to try to become unstuck. The main problem is that after the login, ubuntu crashes, with a screen full of randomly coloured squares. Should I just give up?

    Read the article

  • Xen 4.1 + NVidia driver = Unity has no window decorations

    - by Shade
    I am running Ubuntu 11.10 with Unity. I installed XEN and booted its kernel. The machine booted normally. However, when I started Unity (3D), there were no window decorations. I tried Unity 2D and the window decorations are present there. I then decided to remove the driver (installed through Additional Drivers) and install it manually like this: IGNORE_XEN_PRESENCE=y CC="gcc -DNV_VMAP_4_PRESENT -DNV_SIGNAL_STRUCT_RLIM" ./NVIDIA* ...but the window decorations were still missing in Unity 3D. The window decorations are there when I boot the regular Ubuntu kernel (without XEN support). What could be the problem and how to fix it?

    Read the article

  • Nvidia Driver versions?

    - by Patrick Krenz
    I've looked all over and can't find any reason as to why or how Nvidia names their drivers. for example they have a 330.xxx/340.xxx series that are current but also a 300.xxx and i've found that they aren't always release in order by number. Here's an example on there site with version and release date 331.38 - January 13 334.16 - Feb 7 331.49 - Feb 18 I'm really confused about what driver to actually go with, a few different series versions seem to work adequately and I just want to have an understanding of it and what the best option to work from would be. I really appreciate any information

    Read the article

  • Empty nvidia-settings after install v. 331.67 driver on Ubuntu 14.04?

    - by Victoralm
    I've followed this tutorial (with the following adjustments below) to install v. 331.67 nVidia driver. But get nvidia-settings (NVIDIA X Server Settings) empty... Then, before restart the PC, I make a xorg.conf in the nvidia-settings. The driver make Unity don't start right. So I installed nvidia-331-updates via apt-get and restarted the PC again. After that, the Unity works fine. But the nvidia-settings gets almost empty... Can someone help !?

    Read the article

  • Which nvidia driver from additional drivers option should I choose?

    - by sud_the_devil
    I have NVIDIA Geforece 7025 / nForce 630a integrated gfx card on my UBUNTU 12.04.And I have 4 options in Additional drivers: NVIDIA accelerated graphics driver (version 173) NVIDIA accelerated graphics driver (post-release updates)(version current updates) NVIDIA accelerated graphics driver (version current)[Recommended] NVIDIA accelerated graphics driver (post-release updates)(version 173 updates) I am currently using the First option. My Problem is there I can't update the version 173 to the latest version of drivers using xswat PPA. Whenever I do it and reboot but the NVIDIA xserver sttings always shows me the same version. Which option would allow me to update my drivers to the current version?

    Read the article

  • Unity Not loading after installing Nvidia proprietary drivers

    - by user108705
    I just got a new Nvidia GTX 650. The huge downside at this point is that whenever I use the open source driver, the computer is ridiculously slow. Whenever I use the proprietary drivers and restart, the desktop wallpaper will load and nothing more. I've tried restarting unity only to get a continuous rolling error. With the CCSM options, unity is enabled, yet nothing appears. Whenever I'm using proprietary drivers I get a microcode error which happens so quickly that I can't record the number.

    Read the article

  • Nvidia Fullscreen Metanode "Sliding" Issue

    - by user68202
    i have 2 monitors, the left one is my "main" monitor with 1920x1080_120, the right one my second with 1680x1050_60. (i have a nvidia card setup with twinview) when i play a game or something in fullscreen mode, the full resolution is used in fullscreen (monitor 1 + monitor 2). i read something about the metanodes i can use to shut down the one monitor that i dont need durning a "fullscreen session". i used the following: Option "metamodes" "DFP-0: 1920x1080_120 +0+0, DFP-2: 1680x1050_60 +1920+0; DFP: 1920x1080_120 +0+0, NULL" Its working great, the second (right) monitor is shutting down when i press "CTRL ALT +" and is starting agin when i press the same keystroke. But in the second mode when the second monitor is "down", i got the full "monitor 1 + monitor 2" resolution on my first (left) monitor, i can move my mouse to the right to see the contents of the second monitor and move it again to the left to the what is normally seen on the first monitor. Its something sliding between the 2 monitors on one display. How can i avoid this?

    Read the article

  • Dual-Monitor layout and primary display with Nvidia in Gnome-Shell

    - by java4fun
    I have installed Ubuntu 14.04.1 with gnome-shell version 3.10.4. I can attach a second monitor to my laptop and it is recognized. If I configure it with the nvidia-settings at the right of my laptop then all works fine (the extention of the display works as an extention on the right, not dual). Now what I want is to have the second monitor configured over the laptop-display. When I'm not using gnome-shell, it works correct (I can drag an application from the bottom display to the top normally) but with gnome-shell I have to press Ctrl+Shift+ mouse help, and resize properly the application size. This is for me an enforcing. Is there a way to make that gnome-shell work properly with a bottom/top monitor?

    Read the article

  • TX 1000 Nvidia driver problem

    - by Marduk
    Hello i installed the news ubuntu the other day and istalled all the drivers and then restart the laptop. then i get the login screen with no problem then when i login the black screen in about 10 mins a window popup saying that the application compiz has closed unexpectedly i relunch the app. and then it happens again. i have uninstalled nvidia driver then restarted and the OS works fine.. tryed older driver and still the same problem. anyone out there can help would be great.

    Read the article

  • Ubuntu 12.04 terminal only after nvidia driver upgrade

    - by user1613361
    I recently installed Ubuntu 12.04 (after getting tired of Windows). Installation went fine. I had one problem when I got in suspend mode and then resumed my screen was scrambled. After some googling founds some tips to upgrade my nvidia drivers. I did this, and rebooted but only a terminal windows appears now instead of a graphical desktop. Tried to fix it by several reboots, and removing xorg.conf, but still no luck. Any advice how to fix it and get my desktop back ?

    Read the article

  • How to install a NVIDIA driver for 11.10

    - by Hurumhei
    I have a Thinkpad T61p with Quadro FX570M. I just installed the Ubuntu 11.10. I have tried to download the driver LINUX DISPLAY DRIVER - X86 version 295.20 from NVIDIA, but I do not manage to install it. If I open System Screens, there is an unknovn screen driver. My installation works fine except, fails on resume with a black and white pattern all over the screen and I have difficulties detecting an external monitor. Please, if you have any clue - help me out.

    Read the article

  • Nvidia GTS 450 HDMI problem?

    - by Gercke
    Good afternoon friends. I have a problem with my video card, I have a video card GeForce GTS 450 and a TV Samsung Series 4 43 "and I have it connected to my video card via a cable Mini HDMI to the TV and everything works perfectly, I can instarlar Ubuntu 12.04 (64bits) perfectly, but the problem when I want to change the screen resolution but try to change the driver I install that comes directly from the Nvidia page gave me an error at the end. maximum resolution I presents is 1024x 768 and is supposed to introduce more resolutions and that is connected to the HDMI. A someone else has happened???? Will have some sort of solution???? I hope someone can help me and I thank you very much. Regards

    Read the article

< Previous Page | 5 6 7 8 9 10 11 12 13 14 15 16  | Next Page >