Search Results

Search found 445 results on 18 pages for 'dvi'.

Page 4/18 | < Previous Page | 1 2 3 4 5 6 7 8 9 10 11 12  | Next Page >

  • Dual VGA monitors with Quadro FX 580?

    - by dentrasi
    I have a machine with a Quadro FX 580 card (DVI and two Displayport). Attached to it are two 19" Acer screens, which are both (annoyingly) VGA. The first one works perfectly, with a DVI-VGA adaptor. The second one doesn't work. It's got a VGA cable, which goes into a VGA-DVI converted, which then goes into a DVI-Displayport converter. Initally, I was getting 'Cable Unplugged' on the screen, and it couldn't be seen by Windows or the nVidia control panel. After swapping the VGA-DVI adaptor (which works perfectly on another machine), Windows can now see the monitor. The nVidia panel sees the model and native resolution, but I get a constant 'No Signal' error. Switching to the other Displayport makes no difference. I suspect that the card is seeing a DVI connection plugged into it (nVididaCP shows the monitor as having a DVI connection), and it only sending out a digital signal because of this. Does anyone know of a solution (other than trying to get a Displayport-VGA adaptor), or of a way to force the card to see it as VGA? Thanks, ~Dentrasi

    Read the article

  • Using Toshiba 22EL833 as PC display through HDMI input

    - by Oleg V. Volkov
    I had another Toshiba TV - 19SL738 - connected to this same PC and video card (GTX 8800) through DVI<-HDMI (DVI on PC side, HDMI on TV) before, that was working perfectly at it's native resolution 1360x768. Some time ago I had to change to 22EL833 and immediately faced problem with Windows 7 control panel and NVIDIA control panel both reporting native resolution for new TV as 1080i, 1920x1280, despite TV documentation saying that it have same 1360x768 as previous one. Practical tests confirmed that true native resolution is indeed 1360x768, because plugging in through DVI<-VGA and setting custom resolution through NVIDIA panel shown clear colors and crisp image, while setting anything different with either DVI<-VGA or DVI<-HDMI produced horribly distorted or squished images, with almost unreadable slim lines (as in letters, for example). Now, my problem is that there's no drivers for this TV and I'm unable to get good image while connecting it through DVI<-HDMI directly. The best I've achieved is editing EDID/driver manually, to persuade system that native resolution should be 1360x768, and while image became mostly clear, colors turned to some strange washed out effect, with pools of pure yellow, cyan and magenta there and there filling place of other colors. Gradients also became noticeably stripped as well. Somehow it looks like dithering gone bad and makes me suspect that image is still down/upscaled several times internally somewhere along the line. How can I connect this TV to DVI output of my video card to get best possible clear image, correct colors and correct native resolution?

    Read the article

  • Using Toshiba 22EL833 as PC display with GTX8800

    - by Oleg V. Volkov
    I had another Toshiba TV - 19SL738 - connected to this same PC and video card (GTX 8800) through DVI<-HDMI (DVI on PC side, HDMI on TV) before, that was working perfectly at it's native resolution 1360x768. Some time ago I had to change to 22EL833 and immediately faced problem with Windows 7 control panel and NVIDIA control panel both reporting native resolution for new TV as 1080i, 1920x1280, despite TV documentation saying that it have same 1360x768 as previous one. Practical tests confirmed that true native resolution is indeed 1360x768, because plugging in through DVI<-VGA and setting custom resolution through NVIDIA panel shown clear colors and crisp image, while setting anything different with either DVI<-VGA or DVI<-HDMI produced horribly distorted or squished images, with almost unreadable slim lines (as in letters, for example). Now, my problem is that there's no drivers for this TV and I'm unable to get good image while connecting it through DVI<-HDMI directly. The best I've achieved is editing EDID/driver manually, to persuade system that native resolution should be 1360x768, and while image became mostly clear, colors turned to some strange washed out effect, with pools of pure yellow, cyan and magenta there and there filling place of other colors. Gradients also became noticeably stripped as well. Somehow it looks like dithering gone bad and makes me suspect that image is still down/upscaled several times internally somewhere along the line. How can I connect this TV to DVI output of my video card to get best possible clear image, correct colors and correct native resolution?

    Read the article

  • Cannot increase monitor refresh rate on second monitor

    - by Seany84
    I just purchased an Asus VG248 monitor (supports 144Hz refresh rate) and I can not increase the monitor's refresh rate beyond 60Hz from Windows. My system setup: 2x AMD 6990 graphics cards 1x Asus VG236 connected via DVI cable to graphics card #1 DVI port #1 (120Hz) 1x Asus VG248 connected via DVI cable to graphics card #1 DVI port #2 (60Hz) Does anyone know why I can not set the new monitor to a higher refresh rate?

    Read the article

  • Intel motherboard drops HDMI output after Windows boots

    - by Nick
    I have an Intel motherboard (DH67CL) that uses Intel HD 3000 video. It has two video output ports: HDMI and DVI. Both my monitors are DVI, so I have two cables: a DVI-DVI and an HDMI-DVI. If I plug a monitor into just one of these ports, things work fine. But if I plug monitors into each of the ports, the system startup screens show on both monitors, along with the "Starting Windows..." screen, but as soon as the system boots into Windows, the HDMI-connected monitor goes to sleep. When I use the Intel video control panel, the HDMI-connected monitor is not detected. How do I get a dual-monitor setup to work?

    Read the article

  • Origin of display connector numbers in XServer (e.g. HDMI1, HDMI2, DP1)

    - by Andreas N
    a custom mainboard has a DVI and a DisplayPort connector on the board. Currently, everything that is connected at DVI will be named "HDMI2" in XServer. I can see that by calling the "xrandr" tool (in Ubuntu Trusty Tahr). A display connected to the DP connector will be named "DP1" or "HDMI1", if I use a DP-to-DVI adapter. We are now testing a slightly upgraded board version, which has a newer CPU (Intel J1800, Baytrail) among other things and the position of the DVI and DP connectors are switched. Also, everything at the DVI port is called "HDMI1" and something connected to the DP port gets "DP2" or "HDMI2". Q: What causes these numbers to be produced in this manner and where (probably in the kernel) is it happening? I suspect the cause to be hardware related. Specifically, at which CPU pins the connector pins are routed and attached to. Q: Would it be possible to influence this numbering scheme in order to retain the previous numbering behaviour?

    Read the article

  • Nvidia driver overscan issue second monitor via dvi-d cable

    - by benmichael
    Ok, I know that I have a bit of a bizarre setup, but here goes. I have an old laptop, HP Pavilion 6000. The graphics card in there is a GeForce 7150M. The monitor connection is an old 18pin. The external monitor I use is a Samsung SyncMaster 2333. Don't ask me why, but this monitor only has a dvi-d connection (yes, i have searched it). So I have the monitor plugged into the laptop. If I use any of the Nvidia propriety drivers and try to set the resolution up to 1920x1080 (the monitor's native resolution), I get a massive overscan issue. Over the years I have tried to get this to work, tinkering with my xorg.conf to death. I have also tried this on every Ubuntu since 10.04, on all the corresponding LUbuntus, and on all the Linux Mints since Lisa. Exact same issue. I have even tried it in WinDoze and it works perfectly there (although I did get the error once, but was unable to reproduce it). Using the Open Source drivers it works perfectly iff I switch off the laptop monitor (this makes no difference with the Nvidia drivers). I would have happily gone on using the Open Source drivers, except that since upgrading to LUbuntu 12.10, the Open Source drivers make my monitor completely hazy and have the same overscan issue until I (through the haze, only because I know where things are) go to the monitor settings, activate the laptop's monitor, then deactivate it, and suddenly it comes right. I have to do this every time. So I have to find a way to fix one of them, so I may as well tackle the propriety drivers, hence this overlong question. Amidst other things, I have tried the nvidia-settings, but because it is connected to an 18pin, it detects the monitor as a vga monitor and does not give me overscan correction options. I have tried custom modlines (although there are always more of those try), I have tried using xrandr, and I have tried all the FlatPanelOptions. What I have not tried is a Gentoo build, as I don't have time any more to do that installation, but up to about three years ago when I ran Gentoo exclusively I did not have this issue. Below is an link to an image with a red highlight around the portion of the screen visible to me, the numbers around it are the number of pixels which are cut off. This does seem to drift a few pixels every now and then. Thanks in advance. Nvidia driver issue image

    Read the article

  • Mini DV and DVI Conversion

    - by Kairan
    Video card is: AMD HD 6950 My video card comes with 1xHDMI, 2xDVI, and 2xMini DV ports. I have always used VGA and DVI only. I have the HDMI already connected. I want to hook up 2 more display. One of the display can take a VGA only, the other is an LCD TV so it has VGA, HDMI, Component. I am wondering: 1) What is the best output port to use for the second two monitors? The miniDV or the DVI? 2) I need to run a long cable maybe 15-25 feet, this might change the answer to question 1, is an HDMI long cable 20+ ft going to be better than a 20ft DVI or miniDV ?

    Read the article

  • Is Displayport preferable to DVI for monitor connections?

    - by Sliff
    I'm currently running a single Dell 24" (2408WFP) screen, but am considering adding a second. The problem I've got is that I'm currently using the DVI connector to the on board graphics, so will need to purchase a new graphics card. The 2408WFP also has HDMI and DisplayPort inputs, so I was wondering if DisplayPort is worth considering at present over DVI? On the same subject, does anyone actually manufacture reasonably priced DisplayPort cards - all I've managed to find seem to very expensive workstation cards aimed at financial/design/simulation markets.

    Read the article

  • DVI output to VGA Monitor

    - by Nate
    Am I right in thinking a laptop's female DVI output could be used to plug into a VGA monitor using this adaptor? http://www.nextag.com/Cables-Unlimited-DVI-I-79280721/prices-html?nxtg=d8a0a240511-404B0D198D279591 This would be a 2nd external monitor on a for a laptop's docking station

    Read the article

  • DVI output to VGA Monitor

    - by Nate
    Am I right in thinking a laptop's female DVI output could be used to plug into a VGA monitor using this adaptor? http://www.nextag.com/Cables-Unlimited-DVI-I-79280721/prices-html?nxtg=d8a0a240511-404B0D198D279591 This would be a 2nd external monitor on a for a laptop's docking station

    Read the article

  • nvidia on ubuntu 10.10: switching dvi socket

    - by lurscher
    i have ubuntu 10.10 x86_64 with nvidia 9800 gt and nvidia driver version 270.41.06 my video card has two DVI sockets, but i only use single monitor configuration. Now, i think the main DVI socket might be busted, so i want to try to enable the other as the main one, however, i don't know how to achieve that. I tried just plugging the monitor in that socket but it won't just auto-detect (it would have been way too easy to just work)

    Read the article

  • 3 Monitors powered from a Nvidia fx580

    - by user26067
    I have a NVidia FX580 card in a Dell Precision T1500. The outputs of the back of the card are DVI and 2x displayport. I tried to run 3 screens from the DVI + 2x displayport to dvi adaptors, but the card won't have it... It'll only run 2 of these screens at the same time. http://www.nvidia.com/object/product_quadro_fx_580_us.html says; **Display Support:** Dual Link DVI-I 1 DisplayPort 2 # of Digital Outputs 3 ( 2 out of 3 active at a time ) # of Analog Outputs 1 To me this reads that it will be able to power 2xdvi monitors and 1xvga monitor. Anyone care to confirm or speculate?

    Read the article

  • ATI Firepro v4800 3 monitor support

    - by Jared275
    I have an interesting, yet very frustrating problem. I have two computers, both running win7 32 bit, and both with ATI Firepro V4800 graphics cards. Both are using the DVI port and two DP to DVI adapters to connect 3 monitors. One of them is able to display three desktops, while the other fails at enabling the third, displaying "cannot save changes" in display properties while requiring that one monitor must be disabled when making the change in CCC. I've verified that both computers have the same driver version and that both are using the same DP-DVI adapters. This articles suggests a few things to try, but none of its suggestions seem to work either. I'm kind of at my wits end, hence my posting here. If this is not an appropriate question for SU, I apologize. I admit that I am not very familiar with the differences between dual link and single link DVI, and that is something I have not verified is standardized between the two computers. Is that a possible reason one is working and the other isn't? How do I check if the DVI cable is a single link, or dual link?

    Read the article

  • Ubuntu - connecting 3rd monitor fails with "xrandr: cannot find crtc for output DVI-0"

    - by MDCore
    I've got a laptop with a DVI and VGA output on the back. With everything connected it will only allow me to run 2 of the 3 monitors e.g. laptop display + VGA or DVI+VGA but not all 3. xrandr says I have 2 CRTC's, 0 and 1. The internet says I should be able to share a CRTC if the modeline is the same, and my 2 external monitors are the exact same make and model. How do I convince the software to drive all this hardware?

    Read the article

  • Lenovo t400 laptop docking station with DVI out shows green artifacts when connected to monitor

    - by Roy Rico
    I have a Lenovo T400 laptop, with Windows XP. I have a docking station connected via DVI to my Samnsung monitor. I'm connected at 1920x1200 resolution and I keep seeing all these green dots (when colors are supposed to be near black). It's really bad. I have tried running on my Samsung monitor, at all sizes, even down town 1280x1024 like my monitors at work. When I use the docking station at work, I am able to connect 2 1280x1024 NEC monitors (one via DVI, one through VGA, both through my docking station). I have absolutely no issue there, both displays work great. I have tried to update the drivers, I've even installed the latest video drivers from lenovo's site (DEC-02-2010). Is anyone familiar with this issue? Details: Docking Station: Lenovo Advanced Mini Dock- 250410U - NIB Montior at home: Samsung T240 24-inch monitor Monitors at work: NEC monitors

    Read the article

  • DVI monitor detected only on computer startup

    - by kamil
    I've recently connected a new monitor, LG M2252D-PZ, to a rather outdated computer with Windows XP and Radeon 9600. XP has SP3 installed, video drivers are the latest version back from the times the video card was still supported. My problem is that the monitor works fine only as long as I don't turn it off or switch it to a different input. When I turn it back on, it says "no signal". The key to the problem must be the DVI port, to which the new monitor is connected. The previous monitor was connected to the VGA output, and I've tested that the new one also works fine when connected to the analogue port. Apparently, the computer tests for the presence of a monitor on the DVI port only on startup. The question is, how do I change this?

    Read the article

  • Dual displays not working - NVidia - Ubuntu 12.04 - Second Monitor - Two Screens

    - by user75105
    Graphics Card: NVidia 460 GTX. Driver: NVIDIA accelerated graphics driver (version current) I have one DVI monitor, an old Dell LCD from 2005, and one VGA monitor, an Asus ML238H from 2010 whose HDMI port broke. The Asus is plugged into my graphics card's primary monitor slot and is the better monitor even though it is VGA but my computer defaults to the Dell. This happens when I boot as well; the loading screens, the motherboard brand image, etc. are all displayed on the Dell monitor until Windows loads. Then both monitors work. The same thing happened when I booted up Ubuntu 12.04 but I did not see the second monitor when the log-in screen popped up, nor did I when I logged in. I went to System Settings/Displays and my Asus monitor is not an option. I clicked Detect Displays and the Asus is not detected. I looked at the other questions regarding NVIDIA drivers and recalled my problems with Ubuntu a few years ago and decided to check the driver. I went to Additional Drivers to install the proprietary driver and it looks like it's installed and active but I'm still having this problem. There is another driver option, the post-release NVIDIA driver, but that does not fix the problem either. Also, under System Details/Graphics the graphics device is listed as Unknown, which might indicate that it is using an open source generic driver and not the proprietary NVidia driver. But under Additional Drivers it says that I am using the NVidia driver. Any help is appreciated.

    Read the article

  • Dual displays not working - NVidia - Ubuntu 12.4

    - by user75105
    Graphics Card: NVidia 460 GTX. Driver: NVIDIA accelerated graphics driver (version current) I have one DVI monitor, an old Dell LCD from 2005, and one VGA monitor, an Asus ML238H from 2010 whose HDMI port broke. The Asus is plugged into my graphics card's primary monitor slot and is the better monitor even though it is VGA but my computer defaults to the Dell. This happens when I boot as well; the loading screens, the motherboard brand image, etc. are all displayed on the Dell monitor until Windows loads. Then both monitors work. The same thing happened when I booted up Ubuntu 12.4 but I did not see the second monitor when the log-in screen popped up, nor did I when I logged in. I went to System Settings/Displays and my Asus monitor is not an option. I clicked Detect Displays and the Asus is not detected. I looked at the other questions regarding NVIDIA drivers and recalled my problems with Ubuntu a few years ago and decided to check the driver. I went to Additional Drivers to install the proprietary driver and it looks like it's installed and active but I'm still having this problem. There is another driver option, the post-release NVIDIA driver, but that does not fix the problem either. Also, under System Details/Graphics the graphics device is listed as Unknown, which might indicate that it is using an open source generic driver and not the proprietary NVidia driver. But under Additional Drivers it says that I am using the NVidia driver. Any help is appreciated.

    Read the article

  • Can I use an nVidia Geforce 9600GT for triple-monitor?

    - by John Rudy
    My work PC has (as you can guess from the title) a single Geforce 9600GT card. According to the spec page for this card, both DVI ports are Dual-Link DVI. Both are currently in use on Dell 19" 1280x1024 screens. I'm running Windows 7, with Aero enabled, for what that's worth. What I'm wondering is if I can use a standard DVI splitter cable to drive a third screen from one of the ports -- specifically a Samsung 22" 1680x1050. I'm really outgrowing the two-screen setup, and am strongly considering bringing in my home monitor (which is almost never used) to bolster my usable space. If I can't, then my alternative is to replace one of these Dells with that Samsung, to give me at least a "little" more space. (Before you ask, there's probably no chance I can get my company to spring for a second video card and third monitor. Even if I could, it would not be politically expedient, as a number of my other coworkers are already pretty jealous at my setup.)

    Read the article

  • Switched from DVI to HDMI, possible audio artifacts?

    - by I take Drukqs
    I'm using an ASUS VH236H monitor and an EVGA GeForce 570 GTX both of which are brand new. My monitor has an audio out port for speakers/headphones so I plugged in my headphones and made a random selection from my library when I noticed two things: There are static-like artifacts during "louder" parts of songs. There's what seems to be a volume cap in place. When I crank the volume past 100% in VLC the decibel level does not truly increase but the amount of static does. The cable is not new; I yanked it off of my PS3 when my DVI cable broke. It has been used a good amount on my HDTV and PS3 so I doubt it's a matter of burn-in. I like the way the setup works with an HDMI cable as opposed to DVI because my headphones barely reach my rig whereas I have plenty of slack when they're plugged into my monitor. Thanks in advance for any support. Note: I'm using a high quality HDMI cable from monoprice, AKG K702 headphones, and VLC media player.

    Read the article

  • Graphics card for dual 2560 x 1440 output

    - by Bender
    As the title suggests I want to power two 2560 x 1440 monitors from one graphics card. I am just about to send my second graphics card back to Amazon, so before I purchase a 3rd incorrect card I thought I should get some advice. My latest card has dual DL-DVI output but only supports 2560 x 1440 on one output, the other maxes out at 1920 x 1080. Can graphics cards with two Dual Link DVI adapters (or alternately, one Dual Link DVI and one DisplayPort support the a full 2560 x 1440 on both monitors, or does that depend on the graphics hardware and not only the connectors?

    Read the article

  • External monitor turning black intermittently

    - by coding crow
    I have installed an external monitor (Dell ST2220L, 21.5 inch) on my laptop (Sony Vaio). I am using a DVI-D cable for connection. Since the laptop does not have a DVI-D port, I have connected a DVI-D to HDMI connector on the laptop end and inserted the cable in HDMI port of the laptop.. I have switched off the laptop display and adjusted the screen resolution on the Dell external monitor to 1920 x 1080 and adjusted colors for the Windows 7 and brightness and contrast from the monitor. The problem is the monitor turns of blank intermittently for 1-2 seconds and turns on again at random interval. What could be the reason for this and how to get rid of this problem?

    Read the article

< Previous Page | 1 2 3 4 5 6 7 8 9 10 11 12  | Next Page >