Search Results

Search found 12844 results on 514 pages for 'dual graphics card'.

Page 16/514 | < Previous Page | 12 13 14 15 16 17 18 19 20 21 22 23  | Next Page >

  • Does this video card support sound?

    - by Macros
    Probably a rookie question but here goes...I am looking to buy a new video card for a few year old PC which will be used as a media centre. The card I am looking at is this one http://www.ebuyer.com/product/173708, with the main aim being to play blu-ray films. In the product description it states that the card has 7.1 audio channel support, does this mean it will play the sound from the blu-ray through the HDMI, or do I need a separate sound card?

    Read the article

  • constantly crash on video card

    - by Jason Paddle
    I have this problem with my graphic-card(video card). Video card constantly crash because driver stopped. This is the massage that i get: Display driver stopped responding and has recovered. Display driver NVIDIA Windows Kernel Mode Driver , Version 285.62 stopped responding and has successfully recovered. Video card : MSI GTX470 Twin Frozr Motherboard:Asus P8Z68 V-Pro Tried reinstalling drivers but doesn't help. Any ideas what can be ?

    Read the article

  • Fried graphics card, how to proceed ?

    - by user19496
    Motherboard: Biostar TPower I45 I fried my graphics card (white smoke), by removing the cable marked PCI-E from the card, and then booting. Removed the graphics card, and now the machine is booting, and I can ping it. However I have no possibility to see what is actually happening, because I can't attach a monitor. Can I workaround the lack of monitor in some way, just to see if the motherboard is fine, attach a cable and telnet in or any other way ? Or, do I have to buy and install a new graphics card to be sure ?

    Read the article

  • New graphics card (GTX 760) slowing down entire PC

    - by Cayetano Gonçalves
    My new graphics card is making my PC totally unusable. It boots up really slowly, and when the windows screen comes on, the mouse lags really far behind. Nothing opens at a normal speed. However, when I put in my 5-year old graphics card, it all works fine. I'm currently using a Foxconn Renaissance LGA 1366 Intel X58 ATX Motherboard, Intel Motherboard Intel Core i7-950 Bloomfield 3.06GHz LGA 1366 130W Processor, and a EVGA SuperNOVA 850G2 80PLUS Gold Certified ATX12V/EPS12V 850W Power Supply . I know it can't be the power supply, because I just bought it today to try to fix the problem. I've also installed the newest version of BIOS available for my motherboard. I've also seen extreme variations in CPU while the new graphics card is in, and when the old graphics card is installed, it is much calmer. Any thoughts?

    Read the article

  • Upgrade the Graphics Card for a Dell Dimension 3100

    - by Pat Foran
    Hi, I have a Dell Dimension 3100 Desktop with a 128MB Graphics Card Integrated into the Mother Board. I need to upgrade this 128MB to at least 256MB or 512MB if the system will support same. I am told by Dell that all I have is a PCIx1 slot and that they do not stock a Graphics Card for this. I was told to shop around at Amazon and ebay etc and I would find one there. I have shopped around for some time now and do not know exactly what I am looking for. There are several PCI Graphics Card out there but which one would be the correct one for a Dell Dimension 3100. Can you help me resolve this problem. If you know of a PCIx1 card that will sort out my problem you might please let me have all the details for to purchase it. Regards, Pat,

    Read the article

  • Multiple Graphics cards - Non Multi-Display Setup

    - by Alan Thomas
    Is it possible for me to utilize the processing capabilities of multiple graphics cards, even if I'm only using one monitor? I recently bought a new AMD graphics card, fairly top of the line. I also have a two year old, decent nVidia card. They're obviously very different cards. I don't really mind for gaming because my current card can handle most games fine by itself. I'm concerned about video editing programs such as Adobe Premiere and After Effects. Would the system be able to utilize the power of both cards to, say, render a video? I have both drivers still installed on my machine. And because there is only one monitor connected to my current graphics card (AMD) the nVidia would be connected to no display. So I am wondering whether it would or could be utilized in some way to help in processing video. Thanks!

    Read the article

  • Intel hybrid graphics stuck in power save mode

    - by OZZIE
    I have an Asus UL30V with Intel hybrid graphics. When I bought it it used to switch to turn of the better graphics card to save power and run on a more power effecient one. When I connected the laptop to a power source again I could go to the intel graphics icon on the task bar and click on "switch to enhanced performance mode" or something similar. But like 2-3 months ago it suddenly vanished. I cannot switch modes anymore, it's gotten stuck in power save mode. I can change the windows power setting but that doesn't enable the better graphics card :( What should I do? I've tried restarting the computer and that does not work. I really hope I don't have to reinstall the drivers! It seems like other people have had the same issue: http://en.community.dell.com/support-forums/laptop/f/3518/p/19422695/19994601.aspx , he hasn't even got the same computer as me. I have windows 7 Home Premium x64 Any ideas?

    Read the article

  • Why did my graphics card explode?

    - by C-dizzle
    How did my graphics card "explode" like it did in the picture I have here? And if I was just plugging into my on-board graphics, why wouldn't the PC boot up like it normally would? It wouldn't come up until I opened the cover and removed the card, that's when I found it like this. EDIT: If the card was trying to still operate like this for a while, did it run the risk of damaging the PCI-E slot?

    Read the article

  • Why are graphics cards upside-down?

    - by gbjbaanb
    This is something that has always bugged me - when I install a card into a desktop (ie mini tower) case, the fan is always facing down. Surely, making the card so the components and fan is on the top would help a lot with cooling, allowing those whiney fans to spin a little slower. I know some card manufacturers tried to mitigate this by adding heat pipes and big heatsinks on the back of the card.. but they still put the bits on the same way as everyone else! So, does anyone know why they're all upside-down?

    Read the article

  • Intermittent graphics issue

    - by Rob
    An older (~3 year old) PC I have passed on to my younger brother has a Sony 17" LCD monitor and uses on-board Intel GMA 3100 graphics (supports Aero under Windows 7 and no more is required). Recently he has complained to me twice that after booting up the screen looks all 'fuzzy'. Initially I thought maybe he messed up the ClearType settings by mistake, but resetting that did nothing. The only thing that finally worked was changing the resolution to something else temporarily, then setting it back to the original native (highest supported) resolution. The second time this happened the same 'fix' worked. This is happening only intermittently as of now and is not predictable/reproducible. I suspect either the monitor or on-board graphics is dying. How can I check and confirm which of the two it is (or maybe something else you guys can think of)? If it's the on-board graphics, would I be able to keep using the same motherboard with the addition of a cheap PCI-E graphics card?

    Read the article

  • Monitor has yellow tint when used with the graphics card

    - by artknish
    I have an nVIDIA graphics card when used produces an yellow tint on my monitor's display. I tried with another monitor and it had the same issue. I then switched back to my motherboard's inbuilt graphics output (not sure what the technical term is), and the display seems fine, except I can't get the optimal resolution of 1440x900 to work. So is my graphics card's life over? Or can I get it repaired? Any self remedies without calling a hardware guy? Should I try with a DVI cable? I've been using the VGA cable from my graphics card to the monitor so far. Thanks for your suggestions!

    Read the article

  • Graphics card for dual 2560 x 1440 output

    - by Bender
    As the title suggests I want to power two 2560 x 1440 monitors from one graphics card. I am just about to send my second graphics card back to Amazon, so before I purchase a 3rd incorrect card I thought I should get some advice. My latest card has dual DL-DVI output but only supports 2560 x 1440 on one output, the other maxes out at 1920 x 1080. Can graphics cards with two Dual Link DVI adapters (or alternately, one Dual Link DVI and one DisplayPort support the a full 2560 x 1440 on both monitors, or does that depend on the graphics hardware and not only the connectors?

    Read the article

  • Starting to make 2D games in C++

    - by Ashley
    I'm fairly experienced with C and C#, but I've only ever created console/windows applications. I'm also experienced with AS3 and I've made some flash games. I want to make proper 2D games in C++, but I have no idea where to begin with graphics. There are entire books devoted to game development in C++ that only work with console applications and I'm finding the lack of resources and tutorials for proper 2D games frustrating... I'm also not particularly interested in using existing engines because I want total control of what I create. I've heard of the Allegro library; is it something important/popular that I should look into? How will I use DirectX? Any resources or links to tutorials or information is greatly appreciated.

    Read the article

  • How to make "xrandr" work with GMA500?

    - by Nwbie
    Is it error at driver of graphic chip or Xorg or kernel? I am Asus T91mt with GMA500, Ubuntu 12.04.1. I would like too see only a notice of connection at least. A log of xrandr: $ lspci | grep VGA 00:02.0 VGA compatible controller: Intel Corporation System Corporation System Controller Hub (SCH Poulcbo) Graphics Controller (rev 07) vp@vc:~$ xrandr --verbose xrandr: Failed to get size of gamma for output default Screen 0: minimum 1024 x 600, current 1024 x 600, maximum 1024 x 600 default connected 1024x600+0+0 (0x138) normal (normal) 0mm x 0mm Identifier: 0x137 Timestamp: 26863 Subpixel: unknown Clones: CRTC: 0 CRTCs: 0 Transform: 1.000000 0.000000 0.000000 0.000000 1.000000 0.000000 0.000000 0.000000 1.000000 filter: 1024x600 (0x138) 0.0MHz *current h: width 1024 start 0 end 0 total 1024 skew 0 clock 0.0KHz v: height 600 start 0 end 0 total 600 clock 0.0Hz vp@vc:~$ xrandr --prop xrandr: Failed to get size of gamma for output default Screen 0: minimum 1024 x 600, current 1024 x 600, maximum 1024 x 600 default connected 1024x600+0+0 0mm x 0mm 1024x600 0.0* vp@vc:~$ Please help, i am linux newbie and i am tired ;/

    Read the article

  • Can't get Optimus to work with Ironhide on an Asus N53SN

    - by Musaab
    I installed Ubuntu 11.10 (Same issue in 11.04 btw) and then I installed Ironhide. I went through the configuration, chose the one with the highest confirmation for my system and tested it: > Error: Module nvidia does not exist in /proc/modules > P50 Disabling nVidia Card Succeded (the spelling error is theirs) And it changes nothing. I tried other configurations and got worse results. This has really become a major headache. Any solutions?

    Read the article

  • gnome shell with very high CPU usage

    - by 501 - not implemented
    i'm running ubuntu gnome 13.10 on my dell latiude e6510 with a i5 m560. The I5 comes with a embedded Intel HD 3400 Graphics. The average cpu usage of the gnome-shell is by 160% it's to high, I think. Is there a problem with a driver? If i call the command glxinfo | grep OpenGL it returns: OpenGL vendor string: VMware, Inc. OpenGL renderer string: Gallium 0.4 on llvmpipe (LLVM 3.3, 128 bits) OpenGL version string: 2.1 Mesa 9.2.1 OpenGL shading language version string: 1.30 OpenGL extensions: Greetings

    Read the article

  • How come ASUS sells Ubuntu installed 1225Cs, when we dont have drivers for gma3600?

    - by sumit_gt
    I recently purchased a ASUS 1225C cedar trail netbook. It didn't come preinstalled with Ubuntu like it comes in some markets. I installed Ubuntu 12.04 LTS on it. While Ubuntu works just fine, I am bothered by the lack of proper graphics support because of intel gma 3600. I am aware that this is not a problem of Ubuntu. My question is this: How does ASUS manage to sell these netbooks in other markets with Ubuntu preinstalled? How does ASUS manage to get everything to work? And if they can, why aren't I able to do the same? link: For more on Asus 1225C As you can see, they show Ubuntu working perfectly in the screenshots. Is this false advertising?

    Read the article

  • How to set up multiple GPUs (12.04)?

    - by Brother Erryn
    I have two GPUs: one Intel i915 integrated, and one NVIDIA 560 Ti. This is NOT a hybrid setup, nor a laptop. In Windows 7, each card is connected to a different monitor, with the NVIDIA doing any "heavy lifting". For the life of me I cannot get Ubuntu to recognize the i915, but when logging off or rebooting, the shutdown actually appears on the i915. lshw lists both. I'm running the "current" NVIDIA drivers (not the experimental), but Displays calls its monitor a "Laptop". Is this even possible under Ubuntu? The only things that even looked like potential solutions were for laptop hybrid setups and used Bumblebee, but that doesn't seem to apply here.

    Read the article

  • Gnome 3 freezes on logon on samsung RV 509

    - by Noufal
    I have a Samsung NP-RV509 A0FIN and I tried to install GNU/Linux with gnome 3.2 on it. I tried Fedora 16, Ubuntu 11.10 and Linux Mint 12 RC, but with no success. All of these freezes upon login into gnome shell. I think it is the problem with graphics driver, so I tried xorg-edgers ppa on my last installation, ie., Linux Mint. I also tried various intel graphics packages listed on Synaptic package manager, but no success again. My device configuration is as follows(obtained from windows 7): More details about my computer Component Details Subscore Base score Processor Intel(R) Pentium(R) CPU P6200 @ 2.13GHz 5.6 4.6 Memory (RAM) 4.00 GB 7.2 Graphics Intel(R) HD Graphics 4.6 Gaming graphics 1562 MB Total available graphics memory 5.2 Primary hard disk 12GB Free (50GB Total) 5.9 Windows 7 Ultimate System -------------------------------------------------------------------------------- Manufacturer SAMSUNG ELECTRONICS CO., LTD. Model RV409/RV509/RV709 Total amount of system memory 4.00 GB RAM System type 32-bit operating system Number of processor cores 2 64-bit capable Yes Storage -------------------------------------------------------------------------------- Total size of hard disk(s) 418 GB Disk partition (C:) 12 GB Free (50 GB Total) Media drive (D:) CD/DVD Disk partition (E:) 526 MB Free (191 GB Total) Disk partition (F:) 101 GB Free (177 GB Total) Graphics -------------------------------------------------------------------------------- Display adapter type Intel(R) HD Graphics Total available graphics memory 1562 MB Dedicated graphics memory 64 MB Dedicated system memory 0 MB Shared system memory 1498 MB Display adapter driver version 8.15.10.2202 Primary monitor resolution 1366x768 DirectX version DirectX 10 Network -------------------------------------------------------------------------------- Network Adapter Realtek PCIe GBE Family Controller Network Adapter Broadcom 802.11n Network Adapter Network Adapter Microsoft Virtual WiFi Miniport Adapter Notes -------------------------------------------------------------------------------- The gaming graphics score is based on the primary graphics adapter. If this system has linked or multiple graphics adapters, some software applications may see additional performance benefits. Any help is appreciated, and thanks in advance.

    Read the article

  • Dual DVI and VGA Out

    - by Michael
    I am wanting to add a third monitor with one graphic card. I have been told that a DVI to Dual DVI will display separate views. Right now I do DVI and VGA out. If I get a DVI to Dual DVI out will my VGA still work? Thanks

    Read the article

  • Upgrade only one version of XP to Windows 8 on a dual boot computer

    - by Shane
    I have a computer running Windows XP Pro 32-bit and 64-bit in dual boot. I need to retain Windows XP 32-bit Pro, as I have expensive software that will only run on that specific version. I want to upgrade my 64-bit installation of XP to Windows 8 without losing the 32-bit installation. If I simply use the ISO to upgrade from within my XP 64-bit installation, will I retain dual boot for both XP 32-bit and Windows 8?

    Read the article

  • Problem implementing Blinn–Phong shading model

    - by Joe Hopfgartner
    I did this very simple, perfectly working, implementation of Phong Relflection Model (There is no ambience implemented yet, but that doesn't bother me for now). The functions should be self explaining. /** * Implements the classic Phong illumination Model using a reflected light * vector. */ public class PhongIllumination implements IlluminationModel { @RGBParam(r = 0, g = 0, b = 0) public Vec3 ambient; @RGBParam(r = 1, g = 1, b = 1) public Vec3 diffuse; @RGBParam(r = 1, g = 1, b = 1) public Vec3 specular; @FloatParam(value = 20, min = 1, max = 200.0f) public float shininess; /* * Calculate the intensity of light reflected to the viewer . * * @param P = The surface position expressed in world coordinates. * * @param V = Normalized viewing vector from surface to eye in world * coordinates. * * @param N = Normalized normal vector at surface point in world * coordinates. * * @param surfaceColor = surfaceColor Color of the surface at the current * position. * * @param lights = The active light sources in the scene. * * @return Reflected light intensity I. */ public Vec3 shade(Vec3 P, Vec3 V, Vec3 N, Vec3 surfaceColor, Light lights[]) { Vec3 surfaceColordiffused = Vec3.mul(surfaceColor, diffuse); Vec3 totalintensity = new Vec3(0, 0, 0); for (int i = 0; i < lights.length; i++) { Vec3 L = lights[i].calcDirection(P); N = N.normalize(); V = V.normalize(); Vec3 R = Vec3.reflect(L, N); // reflection vector float diffuseLight = Vec3.dot(N, L); float specularLight = Vec3.dot(V, R); if (diffuseLight > 0) { totalintensity = Vec3.add(Vec3.mul(Vec3.mul( surfaceColordiffused, lights[i].calcIntensity(P)), diffuseLight), totalintensity); if (specularLight > 0) { Vec3 Il = lights[i].calcIntensity(P); Vec3 Ilincident = Vec3.mul(Il, Math.max(0.0f, Vec3 .dot(N, L))); Vec3 intensity = Vec3.mul(Vec3.mul(specular, Ilincident), (float) Math.pow(specularLight, shininess)); totalintensity = Vec3.add(totalintensity, intensity); } } } return totalintensity; } } Now i need to adapt it to become a Blinn-Phong illumination model I used the formulas from hearn and baker, followed pseudocodes and tried to implement it multiple times according to wikipedia articles in several languages but it never worked. I just get no specular reflections or they are so weak and/or are at the wrong place and/or have the wrong color. From the numerous wrong implementations I post some little code that already seems to be wrong. So I calculate my Half Way vector and my new specular light like so: Vec3 H = Vec3.mul(Vec3.add(L.normalize(), V), Vec3.add(L.normalize(), V).length()); float specularLight = Vec3.dot(H, N); With theese little changes it should already work (maby not with correct intensity but basically it should be correct). But the result is wrong. Here are two images. Left how it should render correctly and right how it renders. If i lower the shininess factor you can see a little specular light at the top right: Altough I understand the concept of Phong illumination and also the simplified more performant adaptaion of blinn phong I am trying around for days and just cant get it to work. Any help is appriciated. Edit: I was made aware of an error by this answer, that i am mutiplying by |L+V| instead of dividing by it when calculating H. I changed to deviding doing so: Vec3 H = Vec3.mul(Vec3.add(L.normalize(), V), 1/Vec3.add(L.normalize(), V).length()); Unfortunately this doesnt change much. The results look like this: and if I rise the specular constant and lower the shininess You can see the effects more clearly in a smilar wrong way: However this division just the normalisation. I think I am missing one step. Because the formulas like this just dont make sense to me. If you look at this picture: http://en.wikipedia.org/wiki/File:Blinn-Phong_vectors.svg The projection of H to N is far less than V to R. And if you imagine changing the vector V in the picture the angle is the same when the viewing vector is "on the left side". and becomes more and more different when going to the right. I pesonally would multiply the whole projection by two to become something similiar (and the hole point is to avoid the calculation of R). Altough I didnt read anythinga bout that anywehre i am gonna try this out... Result: The intension of the specular light is far too much (white areas) and the position is still wrong. I think I am messing something else up because teh reflection are just at the wrong place. But what? Edit: Now I read on wikipedia in the notes that the angle of N/H is in fact approximalty half or V/R. To compensate that i should multiply my shineness exponent by 4 rather than my projection. If i do that I end up with this: Far to intense but still one thing. The projection is at the wrong place. Where could i mess up my vectors?

    Read the article

  • Drawing path by Bezier curves

    - by OgreSwamp
    Hello. I have a task - draw smooth curve input: set of points (they added in realtime) current solution: I use each 4 points to draw qubic Bezier curve (1 - strart, 2 and 3rd - control points, 4- end). End point of each curve is start point for the next one. problem: at the curves connection I often have "fracture" (angle) Can you tell me, how to connect my points more smooth? Thanks!

    Read the article

  • Installing/dual-booting Fedora 17 on existing Windows 7 HDD

    - by Moose4
    I have a 64-bit Windows 7 install as the only partition on a 1 TB HDD, with about 350 GB free. I would like to install Fedora 17 as a dual-boot option on this system and give it about 100 GB to play with. If in the Fedora install utility I choose to shrink the W7 partition by 100 GB to give it space, will that cause me to lose my existing W7 data? And how do I go about setting up dual-boot (with Windows 7 as the default)?

    Read the article

< Previous Page | 12 13 14 15 16 17 18 19 20 21 22 23  | Next Page >