Search Results

Search found 12844 results on 514 pages for 'dual graphics card'.

Page 41/514 | < Previous Page | 37 38 39 40 41 42 43 44 45 46 47 48  | Next Page >

  • Can I fully disable my PCIe Video/Graphics Card per BIOS/Software?

    - by Jook
    Because of a quite noisy fan of my HD6700 I was wondering, if I could fully disable my video card through BIOS or even some Software/Windows. Switching to Intel I7 2600 internal video helped already with the noise, but it would be great to have the HD6700 only build in, but not activated/powered. So that the fan could stay compleatly off. Of course, I could just remove the video card, but I would like to avoid that. Is there any way? My Mainboard is an ASUS P8H67-M Pro, with Intel I7 2600 and ATI HD6500

    Read the article

  • Making the most of dual channel DDR2: how do I arrange the sticks?

    - by andrz_001
    I've searched online, but I can't find anything definitive that will put me at ease. I turn to superuser. This is how I have the RAM sticks arranged now: To make the most of the RAM and the dual channel capability, it occurred to me that perhaps I have the sticks arranged incorrectly. Should I move the stick in the DDR2_2 slot one over--to the adjacent, red slot? I appreciate any suggestions. (BTW, can something in BIOS tell me whether I'm running at "optimal" memory speeds??) edit: I'm running Windows XP SP3, 32-bit. Thanks.

    Read the article

  • How do I make YouTube videos fill up the entire screen when using dual monitors?

    - by Jephir
    I am using a dual monitor setup on Ubuntu 9.10 using the TwinView configuration in NIVIDA X Server Settings. My total resolution is 2960x1050 pixels, and my individual monitors are 1680x1050 (primary) and 1280x1024 (secondary). When going into fullscreen mode on any video on YouTube, I only see a cropped version of the video on my primary display as seen below. This does not occur on any other video sharing website - they properly make the video to fill the entire screen on my primary monitor. To my knowledge this problem only happens on YouTube.

    Read the article

  • Can't boot into XP after setting up dual boot with Win 7 vhd.

    - by bebop
    I set up a dual boot for win 7 from vhd on 2 xp machines. One of them everything went fine, and I get the option to choose the os when I turn the machine on. The other was slightly different in that rather than seeing the current OS disk (the one with XP on) as the c:\ when I was setting up windows 7 vhd during install it saw the disk as d:. I didn't think anything of it and went ahead and created a vhd on the d: drive. Now when I turn this machine on, it boots straight to win7 and I never get the option to choose xp. When I look at the boot option in msconfig, I only see Windows 7. How can I go about adding the old XP drive as a boot option at startup again? Edit: Strangely when I rebooted this time (perhaps the first time since I removed the install DVD) it boots to XP. I suppose I'll just have to reinstall windows 7 again in a new vhd...

    Read the article

  • How do I hookup a dual video card Crossfire rig to a single monitor?

    - by Alexed
    I just received a new Alienware box preinstalled with two nice shiny ATI 6950 graphics cards. My goal is to use both of them to give me really good resolution and frame rate on my single monitor. There's zero documentation with the Alienware box and nothing I can find online. Do I just plug the monitor into one of the cards, and expect the two cards to do magic internally? The ATI Crossfire website indicates a need for an "external crossfirex connection cable". Do I need to buy one of these? If so, what kind of cable is it? The AMD Catalyst Control Center shows the second card as "Linked" but when I go to the AMD Overdrive section, the second card is at zero temperature, and clock, suggesting it's not actually doing anything. What do I need to do to make use of both cards? Is it something that doesn't kick in until I do something graphically intensive?

    Read the article

  • How to get full write permission on my second drive after dual boot?

    - by Shaul
    I had a WinXP installation on my hard drive. Then I bought a 2nd drive and installed Win7. Dual boot working fine. But when I'm in the Win7 environment, it appears that I don't have full admin permission on the "D" drive (that's the drive with WinXP). Even though the user I'm logged in as has admin permissions, I have to run my apps with administrator privileges in order to get write access to the D drive. This is not the case when I do stuff on the C drive. I could just get into running those apps that access the D drive in admin mode, but that seems like overkill. Is there some secret switch I can flip so that my D drive acts like my C drive, security-wise?

    Read the article

  • How can I ensure my Samsung Series 7 is actually using the Radeon switchable graphics?

    - by patricksweeney
    I have a Samsung Series 7 with the Radeon 6750M switchable graphics. The ATI software lets you force programs to use the dedicated card. However, I'm not convinced it is actually ever using it, as the frame rates on some non-taxing games (Portal, TF2) are merely OK. To make matters worse, it looks like the ATI Catalyst Control has vanished from my laptop. To make it even worse, you can't download the driver and ATI CCC from ATI's site, you need to download it from Samsung, and the ZIP they provide is corrupted. How can I ensure my Samsung Series 7 is actually using the Radeon switchable graphics?

    Read the article

  • Battery Power when running Ubuntu 14.04 LTS in dual boot

    - by Amro A.
    This is only a general question in order for me to get a better idea of my dual boot (windows 8 & Ubuntu) systems. I noticed that every time I run Ubuntu (which is becoming more often) the battery power gets consumed really fast. I am not performing any special tasks at the moment, just getting to know the system, for example, sound settings, watching videos, surfing the net and so on. When I do the same thing in Windows 8 the battery lives a considerable amount of time longer. Is this something to do with Ubuntu or is it because of the dual boot that I have going on? In other words, if I start running Ubuntu all by itself on my laptop, will it be more power consuming than Windows 8?

    Read the article

  • How do I make YouTube videos fill up an entire screen when using dual monitors?

    - by Jephir
    I am using a dual monitor setup on Ubuntu 9.10 using the TwinView configuration in NIVIDA X Server Settings. My total resolution is 2960x1050 pixels, and my individual monitors are 1680x1050 (primary) and 1280x1024 (secondary). When going into fullscreen mode on any video on YouTube, I only see a cropped version of the video on my primary display as seen below. This does not occur on any other video sharing website - they properly make the video to fill the entire screen on my primary monitor. To my knowledge this problem only happens on YouTube.

    Read the article

  • can I create disk partition for dual-boot Ubuntu on Windows 7 machine without Windows reinstall?

    - by EndangeringSpecies
    I want to setup dual boot Ubuntu on my machine in a separate partition. Plus, ideally, I want to get another, 3rd, partition for further OS experimentation. The hard drive is huge, hundreds of gigs, and essentially unfilled. The machine runs Windows 7 Home. Online I have seen mention of creation of partitions from inside Windows 7. But, I have also heard claims that to create the partition to house Ubuntu Windows has to be reinstalled, frying all the data on the machine. So, which one of these claims are right? Can you create additional partitions for other OS on a big Windows 7 hard drive without reinstall?

    Read the article

  • How to partition a fully used hard drive

    - by MineCraftMan39
    I installed Ubuntu on my laptop which when I got it I had Fedora 13 installed on it as the OS. Now I want to install Fedora 18 in a dual boot with Ubuntu. Problem is when I installed Ubuntu, I didn't partition for dual boot and I gave the entire hard drive to Ubuntu and no longer have the space to dual boot. How can I lower the partition on Ubuntu to make space on the hard drive for Fedora? I want to split the hard drive 50/50 between the two. Thanks in advance.

    Read the article

  • Windows 7 and Ubuntu Boot issue

    - by user115137
    I had the idea to dual boot Win 7 and Ubuntu and what I did was the following: Made a clean install of win 7 using all of my hard drive, next I used the Ubuntu live cd and gparted to partition my drive to be the following: /dev/sda1 ext4 20GB (Linux root) /dev/sda2 ntfs 100GB(Win7) /dev/sda3 ext4 350GB(Home) /dev/sda4 extended 4GB(swap) The thing is, when installing ubuntu I deleted the partition win 7 creates for its boot sector and recovery and then resized the drive to look like what I mentioned, and Ubuntu installed GRUB to the MBR. When GRUB boots I can see Ubuntu but not Windows, how can I chainload it? Or should I fix the windows mbr with the windows 7 installation disk and try to set the dual boot from there? I don't really care which one of the 2 bootloaders I end up using, I just want the dual boot to work out. Thanks

    Read the article

  • OS Isolation: Virtualization or Dual-Boot Duplication, a General How To?

    - by Mr_CryptoPrime
    I want to isolate my windows 7 operating system and I have looked into virtualization. This should work with Linux, however, I do want to still have a way to run windows 7 securely, but without significant performance loss, thus eliminating virtualization for that. I know that you can dual boot because I currently do so with my XP/Linux system. Is there a way that I can duplicate my windows 7 system so I can select one at bootup? This way I can ensure that each OS is isolated and not worry about performance loss. However, I am having a lot of trouble finding a solid method for OS duplication?! Is this even possible or must I buy two versions of win7 and somehow install them separately? Any information regarding this would be helpful, thanks! Essentially I want, Two instances of win7 (not necessarily simultaneously running) Each are isolated from one another so that a security breach in one doesn't affect the other. There is no performance loss in either from doing so

    Read the article

  • Somebody is storing credit card data - how are they doing it?

    - by pygorex1
    Storing credit card information securely and legally is very difficult and should not be attempted. I have no intention of storing credit card data but I'm dying to figure out the following: My credit card info is being stored on a server some where in he tworld. This data is (hopefully) not being stored on a merchant's server, but at some point it needs to be stored to verify and charge the account identified by merchant submitted data. My question is this: if you were tasked with storing credit card data what encryption strategy would you use to secure the data on-disk? From what I can tell submitted credit card info is being checked more or less in real time. I doubt that any encryption key used to secure the data is being entered manually, so decryption is being done on the fly, which implies that the keys themselves are being stored on-disk. How would you secure your data and your keys in an automated system like this?

    Read the article

  • I need to generate credit card surrogates (tokens) that are format preserving.

    - by jammer59
    For an eCommerce application I need to take a credit card and use the real card for passing through to a payment gateway but I need to store, and return to the transaction initiator, a surrogate that is format preserving. Specifically, this means: 1) The number of digits in the surrogate is the same as the real card number (PAN). 2) The issuer type part of the card -- the initial 1,2 or 4 digits remains the same in the surrogate as in the original PAN. 3) The final 4 digits of the surrogate remain the same (for customer service purposes.) 4) The surrogate passes the Luhn mod10 check for a syntactially valid credit card. I can readily handle requirements 1-3 but #4 has me completely stumped! The final implementation will be either t-sql or c#. Any ideas?

    Read the article

  • How to get monitor resolution of 1680x1050 using ATI Radeon HD 2600 PRO in a dual monitor setup

    - by pratikk
    I have Sapphire ATI Radeon HD 2600 PRO graphics card in my Windows Server 2008 x64 machine. It has two outputs a VGA and a DVI. I have connected the DVI to my Dell 24" monitor with 1900x1200 resolution and it works 100%. The VGA I have connected to my second monitor a Samsung 22" with native resolution of 1680x1050. But the ATI driver and Catalyst control centre doesn't show this resolution as an option. If I choose a lower resolution like 1280x1024 it looks really bad and fuzzy. I searched in google and downloaded the powerstrip tool that allowed me to create a custom resolution of 1680x1050 and then this option shows up in Catalyst control centre and my 2nd monitor works fine now. But I don't want to pay for an application to choose a display resolution. Why doesn't ATI show me that option by default even though it has no problem in actually supporting the display at resolution. Is there a way to get 1680x1050 resolution using ATI drivers only ?

    Read the article

  • Display with intel integrated graphics, bitcoin mine with Radeon 6950

    - by karategeek6
    I'm on Ubuntu Linux 11.04 64 bit. I have an intel i5 with integrated graphics and a Radeon 6950, with one monitor. I would like to run my graphics on the integrated card, and run bitcoin mining on the 6950. I have bitcoin mining working when I use the 6950 for both display and mining. Every time I try and and use the integrated graphics instead, OpenCL doesn't recognize my 6950. Using aticonfig --initial when using the integrated graphics for display breaks things. So I used the xorg.conf it created as a basis and tried to manually edit it. I really don't know what I'm doing, though. My last attempt is given below. The graphics ran off the integrated card, but the 6950 wasn't recognized. Any help would be greatly appreciated! xorg.conf: #Section "ServerLayout" # Identifier "Intel Layout" # Screen "Default Screen" # Identifier "aticonfig Layout" # Screen "aticonfig-Screen[0]-0" # Screen 0 "aticonfig-Screen[0]-0" 0 0 #EndSection Section "Module" Load "glx" EndSection # Intel Section "Device" Identifier "Intel Integrated Graphics" Driver "intel" BusID "PCI:0:2:0" EndSection Section "Monitor" Identifier "Default Monitor" Option "VendorName" "Monitor Vendor" Option "ModelName" "Monitor Name" Option "DPMS" "true" EndSection Section "Screen" Identifier "Default Screen" Device "Intel Integrated Graphics" Monitor "Default Monitor" DefaultDepth 24 EndSection # ATI Section "Device" Identifier "aticonfig-Device[0]-0" Driver "fglrx" BusID "PCI:1:0:0" EndSection Section "Monitor" Identifier "aticonfig-Monitor[0]-0" Option "VendorName" "ATI Proprietary Driver" Option "ModelName" "Generic Autodetecting Monitor" Option "DPMS" "true" EndSection Section "Screen" Identifier "aticonfig-Screen[0]-0" Device "aticonfig-Device[0]-0" Monitor "aticonfig-Monitor[0]-0" DefaultDepth 24 SubSection "Display" Viewport 0 0 Depth 24 EndSubSection EndSection

    Read the article

  • Setup dual screen with Catalyst Control Center on notebook so that nothing changes when closing lid (case) ?

    - by Tom
    I have setup a dualscreen setup for my notebook with an ATI video card (ATI mobility radeon hd 5470) using Catalyst Control Center. The monitor extends the desktop of the notebook's screen. This works great. I am running on windows 7 and have set it up to not do anything when I close my laptop's lid (case) other than dim the notebook's display. However, when I do close the lid (notebook's display), the second monitor suddenly changes and I believe it starts behaving as the primary monitor.. I do not want this to happen. So, how do I setup my dualscreen setup so that it won't set the secondary monitor as the primary one when I close my laptop's lid? UPDATE: this video shows exactly what my problem is: http://www.youtube.com/watch?v=Q1ygNHbqI2g

    Read the article

  • How to copy VirtualBox VDI contents to a partition and dual boot the OS from it?

    - by Calmarius
    I'm a Linux user but I keep a compressed Windows XP ISO with me on a pen drive for the case I absolutely need Windows to do something. This works in VirtualBox most of the time. But now I want to play some games, so I would like to run the Windows image natively. My computer don't have CD drive so cannot just burn the ISO and make an install normally. What I trying to do is moving the installed Windows image to a physical NTFS partition on my HDD and set up GRUB to let me dual boot it. I found many tutorials that deal with making VDI to physical drive. But they assume I want to overwrite my entire drive. Moving the raw disk image with dd to the partition resulted in a corrupt partition. I also tried the VMDK trick to use that empty partition and install the Windows on it. Although the text mode phase of the installation finishes without problems, the VM won't work, either crashes and keeps rebooting or just immediately or freezes (depending on how I created the VMDK, with -rawdisk /dev/sda3 or -rawdisk /dev/sda -partition 3).

    Read the article

  • How do I resolve this exercise on C++? [closed]

    - by user40630
    (Card Shuffling and Dealing) Create a program to shuffle and deal a deck of cards. The program should consist of class Card, class DeckOfCards and a driver program. Class Card should provide: a) Data members face and suit of type int. b) A constructor that receives two ints representing the face and suit and uses them to initialize the data members. c) Two static arrays of strings representing the faces and suits. d) A toString function that returns the Card as a string in the form “face of suit.” You can use the + operator to concatenate strings. Class DeckOfCards should contain: a) A vector of Cards named deck to store the Cards. b) An integer currentCard representing the next card to deal. c) A default constructor that initializes the Cards in the deck. The constructor should use vector function push_back to add each Card to the end of the vector after the Card is created and initialized. This should be done for each of the 52 Cards in the deck. d) A shuffle function that shuffles the Cards in the deck. The shuffle algorithm should iterate through the vector of Cards. For each Card, randomly select another Card in the deck and swap the two Cards. e) A dealCard function that returns the next Card object from the deck. f) A moreCards function that returns a bool value indicating whether there are more Cards to deal. The driver program should create a DeckOfCards object, shuffle the cards, then deal the 52 cards. This above is the exercise I'm trying to solve. I'd be very much appreciated if someone could solve it and explain it to me. The main idea of the program is quite simple. What I don't get is how to build the constructor for the class DeckOfCards and how to generate the 52 cards of the deck with different suits and faces. Untill now I've managed to do this: #include <iostream> #include <vector> using namespace std; /* * */ /* a) Data members face and suit of type int. b) A constructor that receives two ints representing the face and suit and uses them to initialize the data members. c) Two static arrays of strings representing the faces and suits. d) A toString function that returns the Card as a string in the form “face of suit.” You can use the + operator to concatenate strings. */ class Card { public: Card(int, int); string toString(); private: int suit, face; static string faceNames[13]; static string suitNames[4]; }; string Card::faceNames[13] = {"Ace","Two","Three","Four","Five","Six","Seven","Eight","Nine","Ten","Queen","Jack","King"}; string Card::suitNames[4] = {"Diamonds","Clubs","Hearts","Spades"}; string Card::toString() { return faceNames[face]+" of "+suitNames[suit]; } Card::Card(int f, int s) :face(f), suit(s) { } /* Class DeckOfCards should contain: a) A vector of Cards named deck to store the Cards. b) An integer currentCard representing the next card to deal. c) A default constructor that initializes the Cards in the deck. The constructor should use vector function push_back to add each Card to the end of the vector after the Card is created and initialized. This should be done for each of the 52 Cards in the deck. d) A shuffle function that shuffles the Cards in the deck. The shuffle algorithm should iterate through the vector of Cards. For each Card, randomly select another Card in the deck and swap the two Cards. e) A dealCard function that returns the next Card object from the deck. f) A moreCards function that returns a bool value indicating whether there are more Cards to deal. */ class DeckOfCards { public: DeckOfCards(); void shuffleCards(); Card dealCard(); bool moreCards(); private: vector<Card> deck(52); int currentCard; }; int main(int argc, char** argv) { return 0; } DeckOfCards::DeckOfCards() { //I'm stuck here I have no idea of what to take out of here. //I still don't fully get the idea of class inside class and that's turning out as a problem. I try to find a way to set the suits and faces members of the class Card but I can't figure out how. for(int i=0; i<deck.size(); i++) { deck[i]//....There is no function to set them. They must be set when initialized. But how?? } } For easier reading: http://pastebin.com/pJeXMH0f

    Read the article

  • EMV credit/debit card chip application

    - by rohitamitpathak
    I am working in smart card industry and familiar with ISO 7816 and Java Card. Till now I worked for ID card, Health card and SIM application using GSM standard. Now I have a chance to work in EMV. Here I have some confusion like: Is a EMV credit/debit card will be a java card containing applet inside it? what standard I need to go through to develop emv debit/credit card application? Is there any good tutorial which helps to understand about EMV debit/credit Card application development? Is there any good simulator to develop and test emv card application?

    Read the article

  • SQL SERVER – Two Puzzles – Answer and Win USD 25 Gift Card

    - by pinaldave
    Today I have two simple T-SQL Puzzle. You can answer them and win USD 25 Gift card. The gift card will be sent in email to winner. You will get choice of Gift Card brand based on your preference and country location. Puzzle 1: What will be the outcome and why? DECLARE @x REAL; SET @x = 9E-40 SELECT @x; The outcome here is obvious as I have used negative number in assignment. What is the reason behind the same? Puzzle 2: Why will be the outcome different from Puzzle 1: DECLARE @y REAL; SET @y = 9E+40 SELECT @y; The outcome of this puzzle very different from puzzle 1  as I have used positive number. There is number six (6) in the resultset why? Msg 232, Level 16, State 2, Line 2 Arithmetic overflow error for type real, value = 90000000000000006000000000000000000000000.000000. How to participate To win the Gift Card USD 25 you will have to answer both of the question on my Facebook page. If you are on twitter – you can increase the chance of winning by tweeting your participation. This contest is open for any one from any country. The winner will be selected Randomly. Winner will be announced on July 7, 2011. Related Post: SQLAuthority News – Monthly list of Puzzles and Solutions on SQLAuthority.com Reference: Pinal Dave (http://blog.SQLAuthority.com) Filed under: Pinal Dave, PostADay, SQL, SQL Authority, SQL Puzzle, SQL Query, SQL Scripts, SQL Server, SQL Tips and Tricks, SQLServer, T SQL, Technology

    Read the article

  • How to Install a Wireless Card in Linux Using Windows Drivers

    - by Justin Garrison
    Linux has come a long way with hardware support, but if you have a wireless card that still does not have native Linux drivers you might be able to get the card working with a Windows driver and ndiswrapper. Using a Windows driver inside of Linux may also give you faster transfer rates or better encryption support depending on your wireless card. If your wireless card is working, it is not recommended to install the Windows driver just for fun because it could cause a conflict with the native Linux driver Latest Features How-To Geek ETC How To Make Hundreds of Complex Photo Edits in Seconds With Photoshop Actions How to Enable User-Specific Wireless Networks in Windows 7 How to Use Google Chrome as Your Default PDF Reader (the Easy Way) How To Remove People and Objects From Photographs In Photoshop Ask How-To Geek: How Can I Monitor My Bandwidth Usage? Internet Explorer 9 RC Now Available: Here’s the Most Interesting New Stuff Smart Taskbar Is a Thumb Friendly Android Task Launcher Comix is an Awesome Comics Archive Viewer for Linux Get the MakeUseOf eBook Guide to Speeding Up Windows for Free Need Tech Support? Call the Star Wars Help Desk! [Video Classic] Reclaim Vertical UI Space by Adding a Toolbar to the Left or Right Side of Firefox Androidify Turns You into an Android-style Avatar

    Read the article

  • XNA - Drawing 2D Primitives (Boxes) and Understanding Matrices in Computer Graphics

    - by MintyAnt
    I have two issues which I wish to solve by creating 2D primitives in XNA. In my game, I wish to have a "debug mode" which will draw a red box around all hitboxes in the game (Red outline, transparent inside). This would allow us to see where the hitboxes are being drawn AND still have the sprite graphics being drawn. I wish to further understand how matrices work within computer graphics. I have a basic theoretical grasp of how they work, but I really just want to apply some of my knowledge or find a good tutorial on it. To do this, I wish to draw my own 2D primitives (With Vertex3's) and apply different transormation matrices to them. I was trying to find a tutorial on drawing primitives using Direct3D, but most tutorials are only for c++, and just tell me to use XNA's Spritebatch. I wish to have more control over my program than just with Spritebatch. Any Help on using Direct3D or any other suggestions would greatly be appreciated. Thank you.

    Read the article

  • Qt Graphics et performances - le coût des commodités, un article de Gunnar traduit par Guillaume Belz

    Le 11 décembre 2009, la documentation de QPainter subissait un énorme ajout concernant l'optimisation de son utilisation. En effet, le bon usage de cet outil n'était pas accessible à tous, il n'était pas présenté dans la documentation. Ceci ne fut qu'un prétexte à une série d'articles sur l'optimisation de QPainter et de Qt Graphics en général. Voici donc le premier article de cette série, les autres sont en préparation : Qt Graphics et performances : ce qui est critique et ce qui ne l'est pas Pensez-vous que cet ajout à la documentation de QPainter sera utile ? Trouvez-vous les performances de vos applications trop faibles ?...

    Read the article

< Previous Page | 37 38 39 40 41 42 43 44 45 46 47 48  | Next Page >