Search Results

Search found 2379 results on 96 pages for 'nvidia prime'.

Page 31/96 | < Previous Page | 27 28 29 30 31 32 33 34 35 36 37 38  | Next Page >

  • Ubuntu 12.04 Full Screen without Unity Hub

    - by Xatolos
    I couldn't find an answer for this, so I thought I would try here. My problem is, I have some games on Wine that play in full screen mode but I can't really get them in full screen mode. When they start up, they end up with the GUI's "HUD" overlapping the game which is really annoying and well kills the option to play the game. This has also happened in a few of my Linux programs as well. While I wrote its a problem with Unity, it really isn't limited to this though. I've done full screen mode on Unity and had the Unity side bar and top menu overlap, I've used Gnome and had the same issue happen, Gnome menu is in the top, Cinnamon had the same issue. Sometimes if I Alt-Tab out of it and back into the program it MIGHT remove the HUD but that doesn't always work, and lets face it, full screen mode that is killed by the GUI just isn't something that can really be ignored. Any suggestions or help would be highly welcomed. My laptop http://www.cnet.com/laptops/asus-g53jw-a1-15/4507-3121_7-34210244.html edit Seems so far to be an issue with Unity and Gnome being in 3D mode, and possibly an issue with my NVidia card and XServe (I think that was it...). Going to Unity/Gnome 2d modes seem to help with this for the moment...

    Read the article

  • Acer aspire v3 771G ubuntu 13.04

    - by Jos
    Gooday, i have this acer and i have alot of boot problems (i suspect windows 8) and now i want to try ubuntu but when i use an usb to "try" ubuntu after the boot i get a black screen. now ive read some of the forums and i found something about NOMODESET i have not tried this as i dont know what this does exactly. now i have found this wiki entry https://wiki.ubuntu.com/Bumblebee , i am by far no programmer and always reading all those commands have always kept me of linux because im scared i will !@#$ things up. is there anyway i can go to NOMODESET in the ubuntu "trial" and can i also include the bumblebee futures (coding?) and in how many ways wil this affect my laptops perfomance? reading the bumblebee entry its seems to be something about nvidia optimus and i dont reallt care much for the power saving, but will it affect any performance? im not a heavy pc gamer but i like tho do some gaming and streaming and such also on a rather big TV in wich this laptop already has it flaws in some games not running properly on 65" if this doesn't work or u advise me not to do this what else can i do to fix windows 8 or either some other linux version? i thankyou in advance

    Read the article

  • Simple Prime Generator in Python

    - by marc lincoln
    Hi, could someone please tell me what I'm doing wrong with this code. It is just printing 'count' anyway. I just want a very simple prime generator (nothing fancy). Thanks a lot. lincoln. import math def main(): count = 3 one = 1 while one == 1: for x in range(2, int(math.sqrt(count) + 1)): if count % x == 0: continue if count % x != 0: print count count += 1

    Read the article

  • How do I force Windows 7 to recognize my Projector?

    - by user63564
    I have a new Dell XPS with an NVIDIA GeForce GT 445M and an old (several years) Epson PowerLite Cinema 550 projector. Windows 7 refuses to recognize that the projector is connected under normal conditions (I'll get to the strange condition in a moment). Here are some things that I have already tried: Confirm that the projector continues to work well on my old Windows XP laptop. Confirm that the video cable (HDMI to HDMI) is connected Make sure the Dell laptop is plugged in to wall power at all times Reboot both the computer and the projector Click "Detect" under the "Connect to an External Display" Windows dialog (no reaction) Click "Rigorous Display Detection" under NVIDIA Control Panel (dialog: none found) Checked "Force Television Detection on startup" under "My display is not shown..." in NVIDIA Control Panel (no effect) Here's where it gets weird... My projector has three states: off, standby and on. Standby means the power switch on the back is on, but the projector is effectively off (no picture, no access to menu or controls). When I plug in the HDMI cable while the projector is in standby, Windows detects the projector! It lets me switch to Duplicate, Extend, or Project Only mode, and adjusts the resolution appropriately. A new Generic Plug-n-Play monitor shows up in my device manager. A "Seiko EPSON PJ" display shows up in my NVIDIA control panel. Then if I turn my projector on, Windows no longer recognizes the display. This is true whether I turn the projector on while the HDMI cable is plugged in, or if I unplug the HDMI cable while turning on the projector. Anyone have any ideas, because I'm completely stumped...?

    Read the article

  • NVIDIA sort sa version 5 de CUDA, simplification de la programmation sur la plateforme de calcul parallèle la plus omniprésente

    NVIDIA sort sa version 5 de CUDA Simplification de la programmation sur la plateforme de calcul parallèle la plus omniprésente NVIDIA a sorti aujourd'hui la version de pré-production NVIDIA CUDA 5, une nouvelle version de la plateforme de calcul parallèle et du modèle de programmation les plus omniprésents, destinée à accélérer les applications scientifiques et d'ingénierie sur les processeurs graphiques. Elle est téléchargeable gratuitement à partir de la zone Développeurs du site Web NVIDIA. Avec plus de 1,5 millions de téléchargements et la prise en charge de plus de 180 applications d'ingénierie, scientifiques et commerciale...

    Read the article

  • Ubuntu 10.10: getting appropiate monitor resolution for lcd hdtv

    - by lurscher
    I'm running Ubuntu 10.10 x86_64 version, with Nvidia 9800 GT, installed 270.41.06 Nvidia drivers following this guide. I have a LG42LH30FR LCD TV connected with the dvi link - RGB PC input I'm able to get 1024x768 resolution without overscan (I can get 1080i = 1366x768 but there is a lot of hidden screen space to the right and I don't know what to do about it). I want to get full HD I can get full HD (1080p = 1920x1080) on Windows XP 64-bit with custom resolution created with Nvidia Control Panel, from reading over xorg.conf configurations it seems I need to add a certain modeling to the monitor configuration, but I don't know where to get the appropriate options for this task any suggestions?

    Read the article

  • Graphics drivers could not find compatible graphics hardware

    - by selak
    I have a desktop with Asus P5KPS/EPU motherboard and Nvidia GF 9600 GT Graphic Adapter. I just set up new windows 7. My problem is that I cannot find a good dirver for my Graphic Adapter. I've downloaded driver from Nvidia.com which say support my GF 9600 GT, but when I installed it, it said This graphics drivers could not find compatible graphics hardware. I go into safe mode, uninstall Standard Graphics Adapter that auto installed then install my Nvidia driver, still not work. Anyone can show me how to solve this problem?

    Read the article

  • How to rotate attached monitor to notebook of Sony's Z-series?

    - by user67175
    Is it possible that Sony is selling a top of the line, hugely expensive computer that does not have the basic ability to rotate an attached monitor? Is it possible that the Z-series simply can't do this? The Windows control panel is missing the normal option for "rotation", as is the Nvidia control panel for "orientation" , no additional rotation software works. Sony sales says they do not know the answer to this. Sony technical supports says that the problem lies with Nvidia, Nvidia technical supports says the problem lies with Sony. Any advice for a fix for this short of returning the computer would be greatly appreciated. Also wondering if this problem is common to computers running Windows 7?

    Read the article

  • Is bundler ready for prime time?

    - by schof
    I'm considering using bundler for deploying a Spree app on Heroku. My question is, is bundler ready for prime time? I know there are some rough edges but I guess I'd like to know more about what the current limitations are and figure out if this is an option for us. Specifically, I'd like to do the git repository stuff git "git://github.com/indirect/rails3-generators.git" gem "rails3-generator Does anyone want to encourage/discourage me from this course of action? Anybody have experience with this on Heroku in particular?

    Read the article

  • Prime Number - Data while loading

    - by Emroot
    Hi, I was trying in Ruby on Rails how to find a prime number. Here is my code : helper : app/helpers/test_helper.rb module TestHelper def prime_number? number index = 2 tmp = 0 while index <= number if tmp < 1 if (number % index) == 0 tmp += 1 end else return false end index += 1 end return true end end and my view : app/views/test/index.html.erb <% (2..100).each do |i| -%> <% if prime_number? i %> <%= i %> <% end -%> <% end -%> So my question is : How can you load data while it's calculating ? I mean if I replace 100 by 100000 in my view, how can I see data on my view while my helper method is calculating ? Do I need to use ajax or rails provide a tool for that ? Thank you.

    Read the article

  • Enumerating large (20-digit) [probable] prime numbers

    - by Paul Baker
    Given A, on the order of 10^20, I'd like to quickly obtain a list of the first few prime numbers greater than A. OK, my needs aren't quite that exact - it's alright if occasionally a composite number ends up on the list. What's the fastest way to enumerate the (probable) primes greater than A? Is there a quicker way than stepping through all of the integers greater than A (other than obvious multiples of say, 2 and 3) and performing a primality test for each of them? If not, and the only method is to test each integer, what primality test should I be using?

    Read the article

  • Simple prime number program - Weird issue with threads C#

    - by Para
    Hi! This is my code: using System; using System.Collections.Generic; using System.Linq; using System.Text; using System.Threading; namespace FirePrime { class Program { static bool[] ThreadsFinished; static bool[] nums; static bool AllThreadsFinished() { bool allThreadsFinished = false; foreach (var threadFinished in ThreadsFinished) { allThreadsFinished &= threadFinished; } return allThreadsFinished; } static bool isPrime(int n) { if (n < 2) { return false; } if (n == 2) { return true; } if (n % 2 == 0) { return false; } int d = 3; while (d * d <= n) { if (n % d == 0) { return false; } d += 2; } return true; } static void MarkPrimes(int startNumber,int stopNumber,int ThreadNr) { for (int j = startNumber; j < stopNumber; j++) nums[j] = isPrime(j); lock (typeof(Program)) { ThreadsFinished[ThreadNr] = true; } } static void Main(string[] args) { int nrNums = 100; int nrThreads = 10; //var threadStartNums = new List<int>(); ThreadsFinished = new bool[nrThreads]; nums = new bool[nrNums]; //var nums = new List<bool>(); nums[0] = false; nums[1] = false; for(int i=2;i<nrNums;i++) nums[i] = true; int interval = (int)(nrNums / nrThreads); //threadStartNums.Add(2); //int aux = firstStartNum; //int i = 2; //while (aux < interval) //{ // aux = interval*i; // i=i+1; // threadStartNums.Add(aux); //} int startNum = 0; for (int i = 0; i < nrThreads; i++) { var _thread = new System.Threading.Thread(() => MarkPrimes(startNum, Math.Min(startNum + interval, nrNums), i)); startNum = startNum + interval; //set the thread to run in the background _thread.IsBackground = true; //start our thread _thread.Start(); } while (!AllThreadsFinished()) { Thread.Sleep(1); } for (int i = 0; i < nrNums; i++) if(nums[i]) Console.WriteLine(i); } } } This should be a pretty simple program that is supposed to find and output the first nrNums prime numbers using nrThreads threads working in parallel. So, I just split nrNums into nrThreads equal chunks (well, the last one won't be equal; if nrThreads doesn't divide by nrNums, it will also contain the remainder, of course). I start nrThreads threads. They all test each number in their respective chunk and see if it is prime or not; they mark everything out in a bool array that keeps a tab on all the primes. The threads all turn a specific element in another boolean array ThreadsFinished to true when they finish. Now the weird part begins: The threads never all end. If I debug, I find that ThreadNr is not what I assign to it in the loop but another value. I guess this is normal since the threads execute afterwards and the counter (the variable i) is already increased by then but I cannot understand how to make the code be right. Can anyone help? Thank you in advance. P.S.: I know the algorithm is not very efficient; I am aiming at a solution using the sieve of Eratosthenes also with x given threads. But for now I can't even get this one to work and I haven't found any examples of any implementations of that algorithm anywhere in a language that I can understand.

    Read the article

  • Wine , macbook pro 5.5, nvidia 9400 and Diablo 3

    - by Jeff Labonte
    I'm using a Macbook pro 5.5 to play Diablo 3 on Ubuntu 12.04LTS 64bit! On Mac OS X it works like a charm but on Ubuntu I'm having some trouble! I have made my research to see what's going wrong! On wine or a VM my graphic has 128mb and on Linux it has 512mb and on Mac OS X 256mb.... well I guess that the graphic card is sharing his memory with the ram! but I've got lags when I play the game! what can I do to have better performances!

    Read the article

  • nvidia optimus laptops

    - by kellogs
    kellogs@kellogs-K52Jc ~ $ lspci 00:00.0 Host bridge: Intel Corporation Core Processor DRAM Controller (rev 18) 00:02.0 VGA compatible controller: Intel Corporation Core Processor Integrated Graphics Controller (rev 18) 00:16.0 Communication controller: Intel Corporation 5 Series/3400 Series Chipset HECI Controller (rev 06) 00:1a.0 USB controller: Intel Corporation 5 Series/3400 Series Chipset USB2 Enhanced Host Controller (rev 06) 00:1b.0 Audio device: Intel Corporation 5 Series/3400 Series Chipset High Definition Audio (rev 06) 00:1c.0 PCI bridge: Intel Corporation 5 Series/3400 Series Chipset PCI Express Root Port 1 (rev 06) 00:1c.1 PCI bridge: Intel Corporation 5 Series/3400 Series Chipset PCI Express Root Port 2 (rev 06) 00:1c.5 PCI bridge: Intel Corporation 5 Series/3400 Series Chipset PCI Express Root Port 6 (rev 06) 00:1d.0 USB controller: Intel Corporation 5 Series/3400 Series Chipset USB2 Enhanced Host Controller (rev 06) 00:1e.0 PCI bridge: Intel Corporation 82801 Mobile PCI Bridge (rev a6) 00:1f.0 ISA bridge: Intel Corporation Mobile 5 Series Chipset LPC Interface Controller (rev 06) 00:1f.2 SATA controller: Intel Corporation 5 Series/3400 Series Chipset 4 port SATA AHCI Controller (rev 06) 00:1f.6 Signal processing controller: Intel Corporation 5 Series/3400 Series Chipset Thermal Subsystem (rev 06) 02:00.0 Network controller: Atheros Communications Inc. AR9285 Wireless Network Adapter (PCI-Express) (rev 01) 03:00.0 System peripheral: JMicron Technology Corp. SD/MMC Host Controller (rev 80) 03:00.2 SD Host controller: JMicron Technology Corp. Standard SD Host Controller (rev 80) 03:00.3 System peripheral: JMicron Technology Corp. MS Host Controller (rev 80) 03:00.4 System peripheral: JMicron Technology Corp. xD Host Controller (rev 80) 03:00.5 Ethernet controller: JMicron Technology Corp. JMC250 PCI Express Gigabit Ethernet Controller (rev 03) ff:00.0 Host bridge: Intel Corporation Core Processor QuickPath Architecture Generic Non-core Registers (rev 05) ff:00.1 Host bridge: Intel Corporation Core Processor QuickPath Architecture System Address Decoder (rev 05) ff:02.0 Host bridge: Intel Corporation Core Processor QPI Link 0 (rev 05) ff:02.1 Host bridge: Intel Corporation Core Processor QPI Physical 0 (rev 05) ff:02.2 Host bridge: Intel Corporation Core Processor Reserved (rev 05) ff:02.3 Host bridge: Intel Corporation Core Processor Reserved (rev 05) kellogs@kellogs-K52Jc ~ $ inxi -SGx System: Host: kellogs-K52Jc Kernel: 3.5.0-17-generic x86_64 (64 bit, gcc: 4.7.2) Desktop: KDE 4.9.5 (Qt 4.8.3) Distro: Linux Mint 14 Nadia Graphics: Card: Intel Core Processor Integrated Graphics Controller bus-ID: 00:02.0 X.Org: 1.13.0 drivers: intel (unloaded: fbdev,vesa) Resolution: [email protected] GLX Renderer: Mesa DRI Intel Ironlake Mobile GLX Version: 2.1 Mesa 9.0.3 Direct Rendering: Yes Manufacturer advertises the K52Jc model which I bought as optimus enabled. However, no traces of it in the output above. Of course, Bumblebee would not start on this machine. Should I rest assured that is a defective / un-optimused machine ? thanks

    Read the article

  • CUDA 3.0 est sorti, avec le support de la nouvelle architecture de NVIDIA, Fermi

    CUDA 3.0 est sorti très récemment, avec le support de la plateforme Fermi, très attendue. Elle n'est pas encore disponible, mais ce n'est plus qu'une affaire de quelques semaines. Cette sortie permet de déjà préparer son code pour la prochaine architecture, tout en bénéficiant d'ores et déjà de grandes améliorations. Citation: Envoyé par Professor Bower, chercheur dans le Quantum ChromoDynamics

    Read the article

  • Ubuntu Unity (64-bit) Bugs on Nvidia dual screen [closed]

    - by Kristofer
    On my work computer I have upgraded 10.04 to 10.10 and now to 11.04. Upgrades have mostly worked well and the dual screen setup is almost working. There are however a couple of annoying bugs which I tried to report but the bug reporting tool told me to ask about the bugs here. Here's what I have found: Auto-hide stops working. It doesn't seem to matter which application is maximised/covering the dock, it just stays on top. I have tested changing the autohide settings with no result. Scenario: One maximised window on each screen. If I now try to drag the window that doesn't have focus by dragging the top of the window nothing happens. I can also not give the window focus by clicking the top bar, I have to click inside the window and then I can drag the window. If I unmaximise the window with focus it does work and I can both change focus and drag the window directly by clicking the top of the maximised window. Any bug fixes coming for these issues? Any work-arounds?

    Read the article

  • How to make Bumblebee work with HP Pavilion DV6T-7000 Quad Edition with Intel HD 4000 and Nvidia GeForce GT 650M 2GB?

    - by user69469
    I just recently bought a HP DV6T-7000 Quad Edition. It has an Intel HD 4000 and a Nvidia GeForce GT 650M 2GB with Optimus. I read that I could use bumblebee to make optimus work, so I installed it. I also installed bumblebee-nvidia and nvidia-current from the ubuntu-x-swat/x-updates ppa. I rebooted, but when I tried to run anything with optirun, the computer would wait ten seconds or so, then do a hard shutdown. I got no log messages from bumblebee, Xorg, or optirun, either. I have purged and reinstalled bumblebee, bumblebee-nvidia, and nvidia-current. I have also set the turned off power management in the bumblebee.conf file to no avail. I am out of ideas about this, and I need both options. Any ideas would be much appreciated.

    Read the article

  • Trying to calculate the 10001st prime number in Java.

    - by user247679
    Greetings. I am doing Problem 7 from Project Euler. What I am supposed to do is calculate the 10001st prime number. (A prime number being a number that is only divisible by itself and one.) Here is my current program: public class Problem7 { public static void main(String args []){ long numberOfPrimes = 0; long number = 2; while (numberOfPrimes < 10001){ if(isPrime(number)){ numberOfPrimes++; } number++; } System.out.println("10001st prime: "+ number); } public static boolean isPrime(long N) { if (N <= 1) return false; else return Prime(N,N-1); } public static boolean Prime(long X,long Y) { if (Y == 1) return true; else if (X % Y == 0) return false; else return Prime(X, Y-1); } } It works okay with finding, say the 100th prime number, but when I enter very large numbers such as 10001 it causes a stack overflow. Does anyone know of a way to fix this? Thanks for reading my problem!

    Read the article

  • Ubuntu 10.04 external monitor problem

    - by Ricky
    I've just installed the 10.04 release on my laptop. I've also installed the nvidia propietary driver as suggested by the system. I've some problem using an external monitor, i'm not able to set the correct resolution using the nvidia-setting utiloty. It doesn't recognize the right resolution of my external monitor. Any advice please? :) I'm a newbie.

    Read the article

  • BSOD - probably video card prob

    - by Shimmy
    I'm receiving the following BSOD: I've tried installing the latest driver from nVidia (296.10), but it says "NVIDIA installer cannot continue", "The graphics driver could not find compatible graphics hardware.": I installed it manually at device manager, but problem still occurs. The problem does not happen in safe mode at all. Note: After restarting the computer when seeing this BSOD, the BIOS welcome screen is full of random black dots spread around: Any solutions? Update: I've tried disabling Aero by following this guide, but that didn't help.

    Read the article

  • Fedora 12 hangs at boot.

    - by Rafael
    Hi all , I'm kinda beginner at linux , i've installed nvidia driver with this command yum install akmod-nvidia every thing went ok , but after reboot the system hangs right after starting atd............[ok] and then i get a cursor " _ " with nothing else happening ! help me please. thanks in advance .

    Read the article

< Previous Page | 27 28 29 30 31 32 33 34 35 36 37 38  | Next Page >