Search Results

Search found 955 results on 39 pages for 'gpu'.

Page 13/39 | < Previous Page | 9 10 11 12 13 14 15 16 17 18 19 20  | Next Page >

  • Monitor randomly shutting down, computer accepting no input, need to restart to get working

    - by Sebastian Lamerichs
    First off, spec list: OS: Windows 7 Ultimate 64-bit SP1 CPU: i7-4820k @ 3.7GHz (stock) GPU: Two 3GB Radeon HD 7970s @ 1.05GHz Mobo: AsRock X79 Extreme6 HDD: 2TB Seagate Barracuda 7200rpm RAM: 16GB quad-channel Kingston 1600MHz PSU: Antec HCG 900W Monitors: Acer S220HQL 1920x1080 + ViewSonic VA2251 1920x1080. Plugged into different GPUs. My problem is that, on a daily-ish basis, my monitors will turn off and not turn back on. My computer will still be running, GPU/CPU/case fans all still going, but the monitors will not turn back on. Additionally, it seems to cease all network activity. It doesn't seem to log any errors at all. I've verified that this is not a monitor issue, as when I press the num/caps/scroll lock buttons on my keyboard, the lights don't change, so the computer is clearly not accepting input. I have noticed a few other people on the internet with this problem, and some have claimed that it was solved by disabling PCI-Express Link State Power Management, but the issue still occurs for me after this. Whilst my CPU and GPUs both run at 100% 24/7, the temperatures are certainly not at dangerous levels, with the CPU averaging 65°C and the GPUs at 70°C and 78°C average. All components are brand new. I have tried forcing MSI Afterburner to start when Windows starts and to force a constant voltage, as this fixed the issue for a few days for another user, but he reported back saying that it had stopped working properly again, so I'm not putting too much faith in this working. Many people have said to adjust display sleep mode settings, but this will clearly not work, as the keyboard lights would still work if the monitors were the issue. The closest I can get to a log file for this issue is the following Folding@Home logs: 14:45:21:WU01:FS00:0x17:Completed 1120000 out of 2000000 steps (56%) 14:46:43:WU00:FS01:0x17:Completed 480000 out of 2000000 steps (24%) 14:46:49:WU01:FS00:0x17:Completed 1140000 out of 2000000 steps (57%) 14:48:30:WU01:FS00:0x17:Completed 1160000 out of 2000000 steps (58%) 14:49:55:WU01:FS00:0x17:Completed 1180000 out of 2000000 steps (59%) As you can see, the second GPU (FS01) stops computation approximately three and a half minutes before the issue occurs (it should be completing 1% every 80-120 seconds), and the first GPU (FS00) continues for a few minutes more before the logs just end. As far as I can tell, the computer has a network failure at the time the first GPU stops working, the latest IRC message I received from this time was at 14:47:58. That being said, there could have just not been any messages between then and 14:50:00, so I'm going to be connecting a laptop to the same bouncer to double-check if it happens again. The GPUs functioned perfectly well in another computer for a significant period of time, so I'm fairly confident that they aren't the issue, which means that this is being caused by either software or the motherboard, or possibly RAM. I really hope it's software. I heard from a forum board that there was a patch from Microsoft that fixed this problem, but "I've forgot which KB it was or the google search terms I used to find the patch, LOL.", so that's not much help. Haven't seen it mentioned by anyone else on about a dozen threads about this issue either. The computer is plugged in via a surge-protected power board, and I've run several other computers and pieces of hardware through it with no issues, so that is not the cause. I have just set the hard disk to never turn off, although I don't believe that that will solve the issue. Strangely, this has only happened when I'm not at the computer (which is actually a minority of the time). Until today it had only happened when I had not been actively using the computer for 6 hours, but today it happened within 10-30 minutes of me last using the computer actively. I have enabled file logging from MSI Afterburner, so hopefully this will shed some light on the issue, but I'm not too optimistic. I've heard that it could be a motherboard problem, but I figured I should ask around before RMAing it. Any help?

    Read the article

  • How do I know if DirectX is using hardware acceleration or software rendering?

    - by JohnIdol
    Is there any DirectX diagnostcs tool which will allow me to understand if Graphics acceleration from my GPU is actually working or software rendering is kicking-in instead? I ask this because If I go properties (right click on desktop) -- settings -- I get an error saying my drivers are not working for my Intel Embedded GPU (Intel Embedded Graphics Driver - IEGD) and the system is defaulting to standard VGA drivers. I am on WinXP Professional.

    Read the article

  • Stopping windows changing colour scheme

    - by dave
    Windows keeps saying my computer is low on memory and changing the aero scheme to basic its really annoying I have 16gb of ram and 2 3gb 6970 gpu's i have no issues at all with anything expect from this my motherboard has a intel hd onboard gpu i think its something to do with this but all my settings are correct. Is there anyway to tell it not to change it without asking me or to tell it how much memory is too low its driving me mad.

    Read the article

  • average temperature of CPU

    - by lego 69
    I downloaded the program named speedFan(for measuring the temperature of the hardware), and I have parameters like these GPU: 60C HDO: 38C Temp1: 50C Core 0: 44C Core 1: 44C Core: 59C Ambient: 0C I know that it shows me temperature of the different parts of my hardware, but I have no idea which ones, can You please explain it and also, is it something wrong with my laptop (because temperature of GPU is 60C) thanks in advance

    Read the article

  • Can a PCI Graphics card, and AGP be used together?

    - by Everyone
    The question pretty much says it all. I use an old 845GBV board ( to-date reliant upon the integrated graphics processor ). All slots on the board are unused. Lately I've been thinking in terms changing it to a dual monitor so that I can use one console for documentation/help/sample code/whatever, the other one to play with code. Assuming this board can handle a PCI GPU, can an AGP4x board coexist with a PCI GPU?

    Read the article

  • Can a PCI Graphics card, and AGP be used together? [closed]

    - by Everyone
    The question pretty much says it all. I use an old 845GBV board ( to-date reliant upon the integrated graphics processor ). All slots on the board are unused. Lately I've been thinking in terms changing it to a dual monitor so that I can use one console for documentation/help/sample code/whatever, the other one to play with code. Assuming this board can handle a PCI GPU, can an AGP4x board coexist with a PCI GPU?

    Read the article

  • How To Make NVIDIA’s Optimus Work on Linux

    - by Chris Hoffman
    Many new laptops come with NVIDIA’s Optimus technology – the laptop includes both a discrete NVIDIA GPU for gaming power and an onboard Intel GPU for power savings. The notebook switches between the two when necessary. However, this isn’t yet well-supported on Linux. Linus Torvalds had some choice words for NVIDIA regarding Optimus not working on Linux, and NVIDIA is now currently working on official support. However, if you have a laptop with Optimus support, you don’t have to wait for NVIDIA — you can use the Bumblebee project’s solution to enable Optimus on Linux today. Image Credit: Jemimus on Flickr How To Create a Customized Windows 7 Installation Disc With Integrated Updates How to Get Pro Features in Windows Home Versions with Third Party Tools HTG Explains: Is ReadyBoost Worth Using?

    Read the article

  • Implementing algorithms via compute shaders vs. pipeline shaders

    - by TravisG
    With the availability of compute shaders for both DirectX and OpenGL it's now possible to implement many algorithms without going through the rasterization pipeline and instead use general purpose computing on the GPU to solve the problem. For some algorithms this seems to become the intuitive canonical solution because they're inherently not rasterization based, and rasterization-based shaders seemed to be a workaround to harness GPU power (simple example: creating a noise texture. No quad needs to be rasterized here). Given an algorithm that can be implemented both ways, are there general (potential) performance benefits over using compute shaders vs. going the normal route? Are there drawbacks that we should watch out for (for example, is there some kind of unusual overhead to switching from/to compute shaders at runtime)? Are there perhaps other benefits or drawbacks to consider when choosing between the two?

    Read the article

  • Wireless driver - how to load manufacturer's STA file (Ralink 3290)

    - by Matt
    Caution: I'm a newb. Hardware: Giada i35G, cedar trail atom with nvidia gf119, railtek ethernet and Ralink 3290 for wireless. Already accomplished: Installed Ubuntu 12.10, loaded GPU drivers and redirected sound out through GPU card to HDMI. Ethernet works like a charm. Issue: Can't get my wireless up and running. There seems to be no package to which I can simply run a sudo get-aspt install ... I found the corresponding Linux driver from the manufacturer's site, but I have not managed to find out what to with the file. Here's the manufacturers site: http://www.ralinktech.com/en/04_support/support.php?sn=501 I get a file with the following name: \2012_0508_RT3290_Linux_STA_v2.6.0.0.bz2 I hope somebody might be able to tell me what to do next. Thanks for reading and apologies for potentially asking a trivial question. Best regards, Matt

    Read the article

  • gpgpu vs. physX for physics simulation

    - by notabene
    Hello First theoretical question. What is better (faster)? Develop your own gpgpu techniques for physics simulation (cloth, fluids, colisions...) or to use PhysX? (If i say develop i mean implement existing algorithms like navier-strokes...) I don't care about what will take more time to develop. What will be faster for end user? As i understand that physx are accelerated through PPU units in gpu, does it mean that physical simulation can run in paralel with rastarization? Are PPUs different units than unified shader units used as vertex/geometry/pixel/gpgpu shader units? And little non-theoretical question: Is physx able to do sofisticated simulation equal to lets say Autodesk's Maya fluid solver? Are there any c++ gpu accelerated physics frameworks to try? (I am interested in both physx and gpgpu, commercial engines are ok too).

    Read the article

  • OpenGL Performance Questions

    - by Daniel
    This subject, as with any optimisation problem, gets hit on a lot, but I just couldn't find what I (think) I want. A lot of tutorials, and even SO questions have similar tips; generally covering: Use GL face culling (the OpenGL function, not the scene logic) Only send 1 matrix to the GPU (projectionModelView combination), therefore decreasing the MVP calculations from per vertex to once per model (as it should be). Use interleaved Vertices Minimize as many GL calls as possible, batch where appropriate And possibly a few/many others. I am (for curiosity reasons) rendering 28 million triangles in my application using several vertex buffers. I have tried all the above techniques (to the best of my knowledge), and received almost no performance change. Whilst I am receiving around 40FPS in my implementation, which is by no means problematic, I am still curious as to where these optimisation 'tips' actually come into use? My CPU is idling around 20-50% during rendering, therefore I assume I am GPU bound for increasing performance. Note: I am looking into gDEBugger at the moment Cross posted at StackOverflow

    Read the article

  • Notebook MSI GT640 - noisy fan - Ubuntu 11.10 x64

    - by pablo
    I have all time this issue when using Ubuntus, the fan (fans?) runs most off the time and between 60%~100% even in idle which is weird. The GPU temperatures are mostly around 80 degrees and I don't think it's right :( I always use Jupiter with power saving preset and add the line in linux default like this: quiet splash pcie_aspm=force i915.i915_enable_rc6=1 So it helps a little (fan just doesn't run all the time 100%...) What else can I do? The notebook spec is: Core i7-720 (8 threads) NVIDIA GTS250M So it's rather one of the first, not so silent and cold notebooks;), with Core i7 family and without optimus technology (as nvidia GPU is the only graphic card built-in) but anyway I don't have theese strange hot times in Windows 7.

    Read the article

  • Coolbits not working

    - by usk
    I want to use coolbits to increase fan speed of my Fermi GPU. 280.13 driver installed. Ubuntu 11.10 I edited /etc/X11/xorg.conf as follows, by pressing Alt+F2 and gksu gedit /etc/X11/xorg.conf I get Section "Device" Identifier "Device0" Driver "nvidia" VendorName "NVIDIA Corporation" BoardName "GeForce GTX 470" Option "NoLogo" "True" Option "Coolbits" "4" EndSection Started getting these messages, Gtk-WARNING **: Unable to locate theme engine in module_path: "pixmap" So I did this, sudo aptitude install gtk2-engines-pixbuf terminal, a@z:~$ nvidia-settings -a [gpu:0]/GPUFanControlState=1 a@z:~$ nvidia-settings -q fans 1 Fan on z:0 [0] z:0[fan:0] (Fan 0) a@z:~$ nvidia-settings -a [fan:0]/GPUCurrentFanSpeed=80 ERROR: Error assigning value 80 to attribute 'GPUCurrentFanSpeed' (z:0[fan:0]) as specified in assignment '[fan:0]/GPUCurrentFanSpeed=80' (Unknown Error). So, it's not working; I can't enter the fan speed percentage. Also from NVidia X Server, there are no fan controls. http://en.gentoo-wiki.com/wiki/Nvidia#Manual_Fan_Control_for_nVIDIA_Settings

    Read the article

  • Blank screen after installing nvidia restricted driver

    - by LaMinifalda
    I have a new machine with a MSI N560GTX Ti Twin Frozr II/OC graphic card and MSI PH67A-C43 (B3) main board. If i install the current nvidia restricted driver and reboot the machine on Natty (64-bit), then i only get a black screen after reboot and my system does not respond. I can´t see the login screen. On nvidia web page i saw that the current driver is 270.41.06. Is that driver used as current driver? Btw, i am an ubuntu/linux beginner and therefore not very familiar with ubuntu. What can i do to solve the black screen problem? EDIT: Setting the nomodeset parameter does not solve the problem. After ubuntu start, first i see the ubuntu logo, then strange pixels and at the end the black screen. HELP! EDIT2: Thank you, but setting the "video=vesa:off gfxpayload=text" parameters do no solve the problem too. Same result as in last edit. HELP. I would like to see Unity. This is my grub: GRUB_DEFAULT=0 GRUB_HIDDEN_TIMEOUT=0 GRUB_HIDDEN_TIMEOUT_QUIET=true GRUB_TIMEOUT=10 GRUB_DISTRIBUTOR=`lsb_release -i -s 2> /dev/null || echo Debian` GRUB_CMDLINE_LINUX_DEFAULT="video=vesa:off gfxpayload=text nomodeset quiet splash" GRUB_CMDLINE_LINUX=" vga=794" EDIT3: I dont know if it is important. If this edit is unnecessary and helpless I will delete it. There are some log files (Xorg.0.log - Xorg.4.log). I dont know how these log files relate to each other. Please, check the errors listed below. In Xorg.1.log I see the following error: [ 20.603] (EE) Failed to initialize GLX extension (ComIatible NVIDIA X driver not found) In Xorg.2.log I see the following error: [ 25.971] (II) Loading /usr/lib/xorg/modules/libfb.so [ 25.971] (**) NVIDIA(0): Depth 24, (--) framebuffer bpp 32 [ 25.971] (==) NVIDIA(0): RGB weight 888 [ 25.971] (==) NVIDIA(0): Default visual is TrueColor [ 25.971] (==) NVIDIA(0): Using gamma correction (1.0, 1.0, 1.0) [ 26.077] (EE) NVIDIA(0): Failed to initialize the NVIDIA GPU at PCI:1:0:0. Please [ 26.078] (EE) NVIDIA(0): check your system's kernel log for additional error [ 26.078] (EE) NVIDIA(0): messages and refer to Chapter 8: Common Problems in the [ 26.078] (EE) NVIDIA(0): README for additional information. [ 26.078] (EE) NVIDIA(0): Failed to initialize the NVIDIA graphics device! [ 26.078] (II) UnloadModule: "nvidia" [ 26.078] (II) Unloading nvidia [ 26.078] (II) UnloadModule: "wfb" [ 26.078] (II) Unloading wfb [ 26.078] (II) UnloadModule: "fb" [ 26.078] (II) Unloading fb [ 26.078] (EE) Screen(s) found, but none have a usable configuration. [ 26.078] Fatal server error: [ 26.078] no screens found [ 26.078] Please consult the The X.Org Found [...] In Xorg.4.log I see the following errors: [ 15.437] (**) NVIDIA(0): Depth 24, (--) framebuffer bpp 32 [ 15.437] (==) NVIDIA(0): RGB weight 888 [ 15.437] (==) NVIDIA(0): Default visual is TrueColor [ 15.437] (==) NVIDIA(0): Using gamma correction (1.0, 1.0, 1.0) [ 15.703] (II) NVIDIA(0): NVIDIA GPU GeForce GTX 560 Ti (GF114) at PCI:1:0:0 (GPU-0) [ 15.703] (--) NVIDIA(0): Memory: 1048576 kBytes [ 15.703] (--) NVIDIA(0): VideoBIOS: 70.24.11.00.00 [ 15.703] (II) NVIDIA(0): Detected PCI Express Link width: 16X [ 15.703] (--) NVIDIA(0): Interlaced video modes are supported on this GPU [ 15.703] (--) NVIDIA(0): Connected display device(s) on GeForce GTX 560 Ti at [ 15.703] (--) NVIDIA(0): PCI:1:0:0 [ 15.703] (--) NVIDIA(0): none [ 15.706] (EE) NVIDIA(0): No display devices found for this X screen. [ 15.943] (II) UnloadModule: "nvidia" [ 15.943] (II) Unloading nvidia [ 15.943] (II) UnloadModule: "wfb" [ 15.943] (II) Unloading wfb [ 15.943] (II) UnloadModule: "fb" [ 15.943] (II) Unloading fb [ 15.943] (EE) Screen(s) found, but none have a usable configuration. [ 15.943] Fatal server error: [ 15.943] no screens found EDIT4 There was a file /etc/X11/xorg.conf. As fossfreedom suggested I executed sudo mv /etc/X11/xorg.conf /etc/X11/xorg.conf.backup However, there is still the black screen after reboot. EDIT5 Neutro's advice (reinstalling the headers) did not solve the problem, too. :-( Any further help is appreciated! EDIT6 I just installed driver 173.xxx. After reboot the system shows me only "Checking battery state". Just for information. I will google the problem, but help is also appreciated! ;-) EDIT7 When using the free driver (Ubuntu says that the free driver is in use and activated), Xorg.0.log shows the following errors: [ 9.267] (II) LoadModule: "nouveau" [ 9.267] (II) Loading /usr/lib/xorg/modules/drivers/nouveau_drv.so [ 9.267] (II) Module nouveau: vendor="X.Org Foundation" [ 9.267] compiled for 1.10.0, module version = 0.0.16 [ 9.267] Module class: X.Org Video Driver [ 9.267] ABI class: X.Org Video Driver, version 10.0 [ 9.267] (II) LoadModule: "nv" [ 9.267] (WW) Warning, couldn't open module nv [ 9.267] (II) UnloadModule: "nv" [ 9.267] (II) Unloading nv [ 9.267] (EE) Failed to load module "nv" (module does not exist, 0) [ 9.267] (II) LoadModule: "vesa" [...] [ 9.399] drmOpenDevice: node name is /dev/dri/card14 [ 9.402] drmOpenDevice: node name is /dev/dri/card15 [ 9.406] (EE) [drm] failed to open device [ 9.406] (II) Loading /usr/lib/xorg/modules/drivers/vesa_drv.so [ 9.406] (WW) Falling back to old probe method for fbdev [ 9.406] (II) Loading sub module "fbdevhw" [ 9.406] (II) LoadModule: "fbdevhw" EDIT8 In the meanwhile i tried to install WIN7 64 bit on my machine. As a result i got a BSOD after installing the nvidia driver. :-) For this reason i sent my new machine back to the hardware reseller. I will inform you as soon as i have a new system. Thank you all for the great help and support. EDIT9 In the meanwhile I have a complete new system with "only" a MSI N460GTX Hawk, but more RAM. The system works perfect. :-) The original N560GTX had a hardware defect. Is is possible to close this question? THX!

    Read the article

  • Tutorial (or livedisk) for multiseat setup with dual-head display supporting openGL direct rendering?

    - by Tobias Kienzler
    I'm currently running Ubuntu 10.10 (64 bit), the GPU is an ATI Radeon HD 4290 onboard a ASUS M4A89GTD PRO/USB3 mainboard. The Ubuntu wiki doesn't cover 10.10 yet and I also don't know if that method would support direct rendering. I heard mentions of xephyr and xgl, what are the differences? Where should I start? I tried the mdm livedisk but that doesn't boot. I'm also willing to try a different distribution if necessary. Edit: 1 Will the HowTo: A well performing, full eye-candy, accelerated pseudo-multiseat setup on a single dualhead GPU. for Ubuntu 9.04 still work? I'm afraid it omits how to setup two input devices and sound however... Edit: 2 http://multiseatonlinux.blogspot.com/2010/06/part-1-setting-up-base.html covers 10.10 but requires pinning gdm. Is that circumvenatable? Also how does that setup have to be modified for dual-head?

    Read the article

  • Ubuntu 12.10 Nvidia GT555M Bumblebee

    - by herczigem
    I have laptop with Nvidia GT 555M video card. System Ubuntu 12.10, kernel Linux 3.5.0-17-generic step what i do: sudo add-apt-repository ppa:bumblebee/stable sudo add-apt-repository ppa:ubuntu-x-swat/x-updates sudo apt-get update sudo apt-get install bumblebee bumblebee-nvidia restart system optirun glxgears This message give me: Cannot access secondary GPU - error: Could not load GPU driver Aborting because fallback start is disabled. Open sudo gedit /etc/bumblebee/bumblebee.conf and change Driver= to Driver=nvidia and KernelDriver=nvidia-current to KernelDriver=nvidia. Restart the system and run optirun glxgears. This message gives me: The Bumblebee daemon has not been started yet or the socket path /var/run/bumblebee.socket was incorrect. Could not connect to bumblebee daemon - is it running? Anybody have idea?!

    Read the article

  • Does an onboard video affect the X windows configuration?

    - by Timothy
    Does the onboard video on the motherboard affect the X windows configuration? My system has onboard and pcie video. The onboard video is a NVIDIA GeForce 7025 GPU, On Board Graphic Max. Memory Share Up to 512MB(Under OS By Turbo Cache). I have a pcie dual head video card installed with two monitors. The video card is a GeForce 8400 GS, with 512mb memory. When installing Ubuntu 12.04, only one monitor worked. When pulling up system settings- Displays it shows a laptop. This is a desktop pc. I did get both monitors to work using nvidia using twinview -- A complicated process! When checking nvidia now it shows the monitors disabled. The Nvidia X server setting does show the GPU and all the information. I was thinking it's seeing the onboard video on the motherboard. Why else would it show laptop?

    Read the article

  • Lenovo Y570 nVidia drivers are giving me problems. (GT 555)

    - by Joe
    So, I've tried the bumblebee "hack" listed HERE:https://github.com/Bumblebee-Project/bbswitch/tree/hack-lenovo I copied every line below into the terminal, is that what I was supposed to do? I'm a Linux noob. git clone http://github.com/Bumblebee-Project/bbswitch.git -b hack-lenovo cd bbswitch mkdir /usr/src/acpi-handle-hack-0.0.1 cp Makefile acpi-handle-hack.c /usr/src/acpi-handle-hack-0.0.1 cp dkms/acpi-handle-hack.conf /usr/src/acpi-handle-hack-0.0.1/dkms.conf dkms add acpi-handle-hack/0.0.1 dkms build acpi-handle-hack/0.0.1 dkms install acpi-handle-hack/0.0.1 echo acpi-handle-hack | sudo tee -a /etc/modules sudo update-initramfs -u (had to use http:// instead of git://, in my uni) It doesn't change anything, I get joe@ubuntu:~$ optirun glxspheres [ 4447.830749] [ERROR]Cannot access secondary GPU - error: Could not load GPU driver [ 4447.830844] [ERROR]Aborting because fallback start is disabled.

    Read the article

  • Drawing of a huge model - How to regain performance?

    - by marc wellman
    I have a huge model I want to draw in my XNA application but because of its size I am experiencing a tremendous loss of performance. The model has about ~50 000 000 edges and has a size on disk of 205 MB in DirectX Format. Please don't ask whether this model has to be that big - yes it has! Is there a way to transfer the model directly to my GPU in order to let the GPU do the drawing like when transferring a VertexBuffer like this: graphicsDevice.Vertices[1].SetSource(_instanceBuffers[i], 0, _sizeofMatrix); because when I try to fill a vertexBuffer with all the vertices I am getting a OutOfMemoryException.

    Read the article

  • What exactly can shaders be used for?

    - by Bane
    I'm not really a 3D person, and I've only used shaders a little in some Three.js examples, and so far I've got an impression that they are only being used for the graphical part of the equation. Although, the (quite cryptic) Wikipedia article and some other sources lead me to believe that they can be used for more than just graphical effects, ie, to program the GPU (Wikipedia). So, the GPU is still a processor, right? With a larger and a different instruction set for easier and faster vector manipulation, but still a processor. Can I use shaders to make regular programs (provided I've got access to the video memory, which is probable)? Edit: regular programs == "Applications", ie create windows/console programs, or at least have some way of drawing things on the screen, maybe even taking user input.

    Read the article

  • Keep getting messages about internal system errors

    - by Tomas Lycken
    I keep getting popups about internal system errors (see screenshot below) on irregular intervals (several times a day), that I don't know what to do about. If I continue through the dialog and try to report the error back to the Ubuntu project, I get a message stating that development on this version of Ubuntu has been completed, and that I should ask for help here if I don't know what to do about it. I don't. If I show the details of the error message, the "executable path" parameter shows /usr/share/apport/apport-gpu-error-intel.py. Is this a bug I should report to Launchpad, or just a configuration error somewhere? If it's a bug, how do I collect the data I (and the devs) need? Update in response to comment: I am running an ASUS N53SN, sporting an Intel Core i7 2630QM CPU and an NVidia GeForce 550M GPU.

    Read the article

  • Samsung a-t-il triché sur les benchmarks pour le Galaxy S4 ? Les performances de l'Exynos 5 remises en cause après des tests

    Samsung : les Galaxy S4 affichent des performances élevées pour les benchmarks Mais moindres pour les autres applicationsTout a débuté avec le post d'un utilisateur mécontent sur le site Beyond3d après qu'il ait réalisé des bancs d'essai pour le GPU de son Galaxy S4. Pendant les essais, l'utilisateur constate que le GPU de son smartphone tourne à 532 Mhz. Cependant, pour toutes les autres applications y compris les jeux (aux résolutions les plus poussées) ce dernier ne tournerait plus qu'à 480 Mhz. « Oh Samsung honte à toi ! » écrira-t-il.Par la suite, Brian Klug et Anand Lal Shimpi, du site web Anandtech, reprennent eux aussi l'expérience de cet utilisateur avec leurs propres smartphones pour en confirmer...

    Read the article

  • Grpahic hardwares

    - by Vanangamudi
    Which vendor provides better GPGPU. my requirements are confined to rendering utilising the GPU for BSDF building for e.g. Intel started providing Ivy Bridge chipset GPU, which are comparably fast to HD5960 cards. I'm not that against nvidia or amd. but I'm a fan of Intel. how it compares to nvidia in price and performance. if possible may I know, how all of them perform with OpenCL?? I'm not sure if it is right to ask it here. but I don't know where to ask.

    Read the article

  • Is it only possible to display 64k vertices on the monitor with 16bit?

    - by Aufziehvogel
    I did the first 3D tutorial over at riemers.net and stumbled upon that my graphic card only supports Shader 2.0 (Reach profile in XNA) which means I can only use Int16 to store the indices (triangle to vertex). This means that I can only store 2^16 = 65536 vertices. Also I read on the internet that you should prefer 16-bit over 32-bit because not all hardware (like mine) does support 32-bit. Yet, I am wondering: Do really all game scenes get along with only so little vertices? I though already faces of people used a lot of polygons (which are made up of vertices?). It’s not relevant for me yet, but I am interested: Do game scenes use only 65536 vertices? Do you use some trade-off to display more (e.g. 64k in GPU buffer rest on RAM) Is there some method to get more into the GPU buffer? I already read on some other posts that there seems to be a limit of 64k per mesh too, so maybe you can compact stuff to meshes?

    Read the article

< Previous Page | 9 10 11 12 13 14 15 16 17 18 19 20  | Next Page >