Search Results

Search found 20049 results on 802 pages for 'virtual drive'.

Page 357/802 | < Previous Page | 353 354 355 356 357 358 359 360 361 362 363 364  | Next Page >

  • Windows 8.1 clean install keeps failing with Oxc0000005 error

    - by joaocandre
    I'm encountering the error Oxc0000005 when making a clean install of Windows 8.1 x64 on a Samsung 840 SSD. I don't know where the error is coming from, I'v already tested several solutions/workarounds I found online and nothing seems to work. The SSD is healthy, I can install Linux (Ubuntu) on it, but apparently the Windows install reports the drive as "locked" when trying to "refresh" the install. According to diskpart, however, the SSD is not in read-only mode, so I don't understand what is going on. I had Windows 8 installed before formatting the drive.

    Read the article

  • Installing 64-bit Ubuntu alongside 32-bit Ubuntu?

    - by Macha
    I have a 64-bit processor in my PC, but because of worries over application compatibility, up until now I have been using 32-bit Ubuntu (and 32-bit Vista because Dell wouldn't sell me 64-bit with my PC). Is it possible for me to install 64-bit Ubuntu alongside 32-bit ubuntu and 32-bit Windows Vista, so I can choose between them at boot and share data, and without uninstalling my 32-bit Ubuntu? My partitions are as follows Drive 1: 10 GB Vista recovery partition (E:), 240 GB Windows NTFS parition (230 GB used, C:). Drive 2: 167 GB Windows NTFS Partition (130 GB used, D: ), 8 GB swap partition, 13 GB / partition (6 GB used), 62 GB /home partition (20 GB used).

    Read the article

  • Can you use a USB dongle inside a VMWare ESX Virual Machine?

    - by Keith Sirmons
    Howdy, I need to know if a USB dongle that is required as a license key for a piece of software will accessible from the physical host machine. This will be a small vSphere 4 installation targeting the quick backup and system restore capabilities of VMWare, not specifically HA, so I am not to worried about the virtual machine automatically failing over to another physical machine and the dongle not being accessible. Does ESX have the capability to map a physical USB port or device to a specific Virtual Machine? I believe this is the dongle: Sentinal Superpro USB Dongle by Rainbow Technologies Thank you, Keith

    Read the article

  • How long should diskpart take?

    - by sam
    I am using diskpart to extend a drive that is actually a VHD. I've already extended the VHD. It's on Windows 2003 and the C drive doesn't contain the swap file and the available space is contiguous. However didn't see the Note about the Resource Kit diskpart for download is not for Windows 2003. So I did the extend using the Windows 2000 version. Not sure if this is the reason but Diskpart is sitting there now for about 15 minutes or so and it's only gotta extend by 10GB. Should it be taking this long? Am I asking for trouble now that I've used a Windows 2000 version of diskpart on a Windows 2003 machine (VM)?

    Read the article

  • File History - Unable to scan user libraries for changes and perform backup of modified files for configuration

    - by azl
    When trying to run the File History tool in Windows 8 it runs for about 2 seconds then stops. No files are backed up to the selected drive. In the event viewer the only error that appears is: Unable to scan user libraries for changes and perform backup of modified files for configuration C:\Users\win8User\AppData\Local\Microsoft\Windows\FileHistory\Configuration\Config I've tried deleting both the configuration files and the FileHistory directory on the target drive. Setting up File History again results in the same error. Is there a better way to track down what is causing the failure? Or somehow get the File History tool to create a more verbose log file that shows what is causing the problem?

    Read the article

  • 'inode table usage' spiking every morning at 8am

    - by Harry Wood
    I installed munin on my ubuntu server. It's showing my 'inode table usage' spiking every morning at 8am. It then rapidly curves down and settles over the course of several hours. What might cause this? I thought it might be something running in /etc/cron.daily but this was set to run at 6a.m. and I've changed it to 4am. The spike remains at 8am. I also enabled cron logging, but can't see anything getting launched at 8a.m. It's a virtual server hosted by memset. Could it be caused by something happening on the virtual host?

    Read the article

  • Why does httpd handle requests for wrong hostnames in SSL mode?

    - by Manuel
    I have an SSL-enabled virtual host for my sites at example.com:10443 Listen 10443 <VirtualHost _default_:10443> ServerName example.com:10443 ServerAdmin [email protected] ErrorLog "/var/log/httpd/error_log" TransferLog "/var/log/httpd/access_log" SSLEngine on SSLProtocol all -SSLv2 SSLCipherSuite HIGH:MEDIUM:!aNULL:!MD5 SSLCertificateFile "/etc/ssl/private/example.com.crt" SSLCertificateKeyFile "/etc/ssl/private/example.com.key" SSLCertificateChainFile "/etc/ssl/private/sub.class1.server.ca.pem" SSLCACertificateFile "/etc/ssl/private/StartCom.pem" </VirtualHost> Browsing to https://example.com:10443/ works as expected. However, also browsing to https://subdomain.example.com:10443/ (with DNS set) shows me the same pages (after SSL certificate warning). I would have expected the directive ServerName example.com:10443 to reject all connection attempts to other server names. How can I tell the virtual host not to serve requests for URLs other than the top-level one?

    Read the article

  • activesync not working with forms based authentication

    - by Chad
    I have an Exchange 2003 SP2 single backend server with an SSL cert. I was having trouble getting OMA to work so I found a MS article about making a reg hack and creating a new Exchange-OMA virtual directory. I am able to connect and access content from my mailbox by using secure mail.domainname.com/oma and using my credentials. ActiveSync was not working on a Windows mobile phone or iPhone. I found another article about using Forms Based Authenication and SSL on a single Exchange server environment and the fix was to elliminate FBA and SSL for the Exchange virtual directory. That allows ActiveSync to now work. I have very few mobile users, but they are management, so I need to make ActiveSync work but I would like to get back to using SSL. http://support.microsoft.com/kb/817379 Any ideas about this setup? Thanks.

    Read the article

  • Windows Server 2008 R2 - What now?

    - by MrStatic
    So I installed Windows Server 2008 R2, I have Exchange 2010 running for some mail. I installed BlackBerry Enterprise Server Express for kicks. This server runs dns/dhcp for 2 laptops. But what now? It has 200g's free on the main drive and a 150g external that I have mapped as a network drive. I am not sure what else to do with this thing. It was my main station but since my laptop is nicer I am running it headless and using the lcd monitor for it as a second monitor for the laptop when I am at home.

    Read the article

  • Moving Windows XP from ICH10R RAID 5 to single disk using Linux [migrated]

    - by tudor
    A friend's machine running Windows XP refused to boot recently which is running 3 SATA disks on RAID 5 (which was previously upgraded from RAID 1 not by me). I have determined there to be a disk failure. The disks have been replaced many times in the past few years. I wish to backup the RAID5 partition before I try anything to fix it. The RAID chipset used is ICH10R/DO. So, I plugged in an extra IDE drive and an Ubuntu USB key and looked at the RAID. The partitioning is a mess, but I did find at least one degraded but working RAID array with two partitions, one 79GB and the other 86GB. Then I: 1) Partitioned my IDE disk using fdisk to have a partition of 80GB and bootable, and marked as NTFS. 2) dd the contents of the array to the partition 3) disconnected everything else 4) inserted a Windows XP CD and ran fixboot, fixmbr, and bootcfg. They all run ok and claim that they worked. (e.g. bootcfg detects the Windows partition, fixboot returns saying that it was written correctly.) However, I'm still getting an error like "DISK FAILURE, BOOT DISK NOT FOUND". I have tried running the GRUB rescue disk, which also runs ok, but won't boot into Windows. It just stops with a flashing cursor after chainloader +1, boot. One clue may be that the partitions appear to be wack. One disk has a 79GB RAID partition on a 500GB drive with a offset, the second disk has a 320GB RAID partition across the whole drive. Additionally, the BIOS lists the RAID size as being 149GB. I don't see how this works. How are they even assembling the array when the partitions are so different? I have also tried running the Windows XP automated repair tool, but that didn't work either. I'm presuming this is something simple. Perhaps Windows is attempting to boot into RAID and, upon not finding it, simply crashing? Perhaps the 79GB partitions offset means that it's looking into the disk by that much? Please help!! To clarify: I want to make the single IDE disk bootable with a copy of the array so that I can prove/disprove that it's just that Windows has become corrupted, and use windows tools to correct it before attempting the same thing on the RAID array. That way I have a working backup and can show the process I used to fix it.

    Read the article

  • Critique My Backup and Storage Plan

    - by MetaHyperBolic
    My current storage (RAID-1 off of a hardware RAID card) and backup (a spare drive) solutions for my home network are inadequate. I have too much data scattered on various one-off drives. It is time to evolve. Backups seem simple enough, at least: lots of big drives. However, I am bewildered by the number of choices for small home storage. The Drobo S looks appealing. So does the ReadyNAS. I am not looking for bunches of shiny features, I'm mostly interested in reliability. I am not interested in building Yet Another PC to create a file server or doing something in the cloud, or whatever. I'm stupid, so I am keeping it simple. Requirements for Main Volume: Starting working space roughly 2TB, with options for growth up to 5TB RAID or something RAID-like with at least one parity drive eSATA II for speed during backups Ability to shut down gracefully when alerted of low power by a UPS Optional but Desirable: Will take 2TB drives now with options for the larger 3TB drives coming in 2010-2011 Optional but Desirable: : RAID-6 or something similar, with two parity drives Optional but Desirable: : Hot spare Ethernet connection not required, as the volume will be shared via the same machines which runs my home print server Backups: Backup performed via ROBOCOPY in mirror mode to an external hard drive via a eSATA II connection. Start with rotating between two external 2TB hard drives, will go up to six external 2TB drives. Start with a weekly backup, move to a bi-weekly backup as more drives are added. Move to 3TB drives as the size of my main volume increases. Backup drives will be stored on an off-site location. Hard drives: I plan on buying all of the same model, but different batches from different vendors. I found a "burn-in" utility with which I can pound away on the drives for a couple of weeks before adding them to the backup pool or the main volume. I estimate that I am looking at roughly $1,500 to start, once I start throwing in two TB drives for backup and four for storage. So, are there any obvious flaws in my plan? What have I overlooked? Any suggestions for the storage device for my main volume that fits my requirements? Or do I just keep it simple, 2 drives in RAID-1, then perform due diligence with my backups, accepting that I will have to buy a whole new unit when my data grows past 2TB?

    Read the article

  • Software RAID 10 on Linux

    - by vpetersson
    For a long time, I've been thinking about switching to RAID 10 on a few servers. Now that Ubuntu 10.04 LTS is live, it's time for an upgrade. The servers I'm using are HP Proliant ML115 (very good value). It has four internal 3.5" slots. I'm currently using one drive for the system and a RAID5 array (software) for the remaining three disks. The problem is that this creates a single-point-of-failure on the boot drive. Hence I'd like to switch to a RAID10 array, as it would give me both better I/O performance and more reliability. The problem is only that good controller cards that supports RAID10 (such as 3Ware) cost almost as much as the server itself. Moreover software-RAID10 does not seem to work very well with Grub. What is your advice? Should I just keep running RAID5? Have anyone been able to successfully install a software RAID10 without boot issues?

    Read the article

  • Suggestions for sharing and using data between Ubuntu and Windows 7 dual boot

    - by wizjany
    Note: TL;DR, scroll to bottom for summary. I recently set up my computer for a dual boot between Ubuntu 9.10 and Windows 7. My current drive setup is as follows. | A1 | A2 | | B1 | B2 | B3 | A1: 100 mb, windows 7 "System Reserved" boot partition A2: 230 gb, data section, this one needs to be shared between the operating systems B1: 125 gb, windows 7 OS B2: 123 gb, ubuntu OS B3: 2gb, linux swap space Pretty much i want to have my documents, music, pictures, videos, etc accessible from both operating systems. My first attempt involved making the data (a2) partition NTFS, and moving my home folder from ubuntu to the data partition. However, as I read NTFS does not work nice with permission, and it messed up my home folder. My next idea is one of the following: 1) format the data partition to ext2/3/4 and move my home folder from linux there, and get a driver to read ext partitions in windows 7. The problem with that is that most of the ext drivers/software are not compatible with windows 7 or do not integrate with windows explorer (I really don't want to open a separate software window just to access my data, plus it's probably not compatible with other software.) http://www.fs-driver.org/ looks promising, but I'm not sure how it works with ext4 and windows 7 (not officially supported, when trying in vista compatibility mode, it tells me I need to format the ext drive to use it). My next idea, 2) keep the home folder in ubuntu where it is, but create symlinks for the Documents, Music, etc folders to an NTFS formatted Data (A2) partition, and add those locations to the windows 7 libraries. I'm not totally sure how the permissions would work out, but it should be fine since it's only the documents, music, etc and not the important config files in the rest of /home/user/. Correct me if I'm wrong. Currently, symlinks is my best idea, although i'm not sure how it will work. Any suggestions, additions to my ideas, links, pointers, whatever would be greatly appreciated. Even if it means i should reformat both my drives and repartition (2 250gb drives if you want to suggest a setup for that), I won't be too opposed if that's the best suggestion (I've gone through the format/install/format/reinstall process 5 times over the past 3 days, once more won't hurt me). TL;DR, summary: I have two hard drives. One is partitioned for Ubuntu and Windows 7, the second one I want accessible to both operating systems to store documents, music, pictures, videos, etc. Suggestions on how to set up the data drive please =) P.S. bonus if I can get an apache server document root folder working between the two OS's as well (permissions could become very complicated, so don't worry too much about that) P.P.S. Related question, but data viewing is one way: http://superuser.com/questions/84586/partition-scheme-and-size-for-dual-boot-windows-7-and-ubuntu-9-10-with-separate-p

    Read the article

  • Is it possible to fake a Windows install for grub to add to boot menu?

    - by Mussnoon
    When doing a fresh install of a Linux distro (Ubuntu, for instance) on a fresh hard drive, if I want to install Linux first, and Windows later, is it possible to make grub think there's a Windows install on the first partition so that it'll be added to the boot menu after the installation is complete? To illustrate, I have a new hard drive and have created two primary partitions (both still raw) and two logical (Ext4 and Swap). I want to install Ubuntu on the Ext4 partition first, and some version of Windows on the first primary partition only after that (because I currently don't have a Windows install disk, but do have one for Ubuntu). Is it possible to make Ubuntu add an entry for Windows right now and avoid having to repair grub after I've installed Windows?

    Read the article

  • Good Linux Distro for Disconnected Netbook

    - by MrWizard54
    I'm deployed to Afghanistan with the Army and I have a friend who's netbook had his hard drive take a dump on him. He ordered a new hard drive and I was able to download and burn a copy of Ubuntu to disk from work. However the default install doesn't support most of the media that he wants to watch (AVI files, probably some DIVX video) without installing extra packages. We don't have internet in the tent and really don't have a vaiable option for downloading additional packages through the package manager anywhere here. The computer is a small HP netbook. All my friend wants is to watch ripped movies. Does anyone know of a good way to do either of the following: Download packages seperately and install them via CD A distro that is going to come preloaded with all of the packages and needed to watch just about any type of video file you can think of? Thanks in advance, Andy

    Read the article

  • IPvsadm not equally balancing on wlc scheduler

    - by davidsmalley
    For some reason, ipvsadm does not seem to be equally balancing the connections between my real servers when using the wlc or lc schedulers. One real server gets absolutely hammered with requests while the others receive relatively few connections. My ldirectord.cf file looks like this: quiescent = yes autoreload = yes checktimeout = 10 checkinterval = 10 # *.site.com http virtual = 111.111.111.111:http real = 10.10.10.1:http ipip 10 real = 10.10.10.2:http ipip 10 real = 10.10.10.3:http ipip 10 real = 10.10.10.4:http ipip 10 real = 10.10.10.5:http ipip 10 scheduler = lc protocol = tcp service = http checktype = negotiate request = "/lb" receive = "Up and running" virtualhost = "site.com" fallback = 127.0.0.1:http The weird thing that I think may be causing the problem (but I'm really not sure) is that ipvsadm doesn't seem to be tracking active connections properly, they all appear as inactive connections IP Virtual Server version 1.2.1 (size=4096) Prot LocalAddress:Port Scheduler Flags -> RemoteAddress:Port Forward Weight ActiveConn InActConn TCP 111.111.111.111:http lc -> 10.10.10.1:http Tunnel 10 0 10 -> 10.10.10.2:http Tunnel 10 0 18 -> 10.10.10.3:http Tunnel 10 0 3 -> 10.10.10.4:http Tunnel 10 0 10 -> 10.10.10.5:http Tunnel 10 0 5 If I do ipvsadm -Lnc then I see lots of connections but only ever in ESTABLISHED & FIN_WAIT states. I was using ldirectord previously on a Gentoo based load balancer and the activeconn used to be accurate, since moving to Ubuntu 10.4 LTS something seems to be different. # ipvsadm -v ipvsadm v1.25 2008/5/15 (compiled with popt and IPVS v1.2.1) So, is ipvsadm not tracking active connections properly and thus making load balancing work incorrectly and if so, how do I get it to work properly again? Edit: It gets weirder, if I cat /proc/net/ip_vs then it looks like the correct activeconns are there IP Virtual Server version 1.2.1 (size=4096) Prot LocalAddress:Port Scheduler Flags -> RemoteAddress:Port Forward Weight ActiveConn InActConn TCP B86A9732:0050 rr -> 0AB42453:0050 Tunnel 10 1 24 -> 0AB4321D:0050 Tunnel 10 0 23 -> 0AB426B2:0050 Tunnel 10 2 25 -> 0AB4244C:0050 Tunnel 10 2 22 -> 0AB42024:0050 Tunnel 10 2 23

    Read the article

  • Can't connect disk management to remote XP PC from local Win7/Server 2008

    - by Grez
    Scenario as follows: support technicians using Windows 7 PC's or Server 2008 terminal server are unable to connect Disk Management MMC snap in to a remote PC when the remote device is running Windows XP. "Disk management could not start Virtual Disk Service (VDS) on ". This can happen if the remote computer does not support VDS, or if a connection cannot be established because it was blocked by Windows Firewall." Connecting from another XP machine or 2003 server to the same XP machines works fine. Even connecting from XP/2003 to the Win7 or 2008 server works fine. Windows firewall disabled on all devices. I'm guessing this is something to do with the fact that XP uses logical disk manager service whereas Win7/2008 use Virtual disk manager service. But there doesn't seem to be any way to use logical disk manager service from 7/2008 to connect to XP...

    Read the article

  • kernel: journal commit I/O error

    - by jasondewitt
    I am having some problems with a Dell 1950 server. I am installing RHEL 4.6 along with Oracle and some other software on here. I am randomly getting an error message saying "kernel: journal commit I/O error" on my ssh session and on the monitor I have hooked up to the server I see an error scrolling by that says "EXT3-fs error (device sda5) in start_transaction: Journal has aborted." It has happened several times but never at the same point during the install. Actually, this last time the system was up and running and I was just trying to import a database into oracle. This has happened on several hard drives, so I'm pretty sure that is not the problem. This makes me think the raid controller is going bad. What do you guys think? ** UPDATE ** Pretty sure it was a bad hard drive. I threw another drive in the server and it's been running for about 48 hours with out problems.

    Read the article

  • Mac Book Refuses to Boot

    - by pas09
    I have a Macbook that has unfortunately died on me. I randomly got a pop up message that said I needed to restart my computer, and once I did, I was greeted with the blinking folder question mark start up screen. I've tried everything, including running Disk Utility repair and reformatting the hard drive and reinstalling OS X. All of my data is backed up and safe, I just need my computer to start again. Before I go off and buy a new hard drive, I wanted to see if there is anything I might have forgotten.

    Read the article

  • SQL 2008 Reporting Services Report Manager blank on Win 7

    - by Daniel Root
    I have installed SQL Server 2008 Reporting Services on my development laptop running windows 7. When I browse to the report manager, only the header is shown. The content pane is blank, and there is no 'Site Settings' option, New Folder options, etc. I found plenty of similar stories about SQL 2005, and then the fix was to manage permissions in the virtual directory in IIS. However, in 2008, there is no virtual directory - RS manages this outside of IIS. Per other serverfault articles, I've tried the following: adding my account to the reporting services group changing the service to run as local system made sure localhost was trusted site in IE ran IE as admin installed SQL Server 2008 SP1 There are no errors in the event viewer. Perhaps related, if I browse to the webservice, I also get the error: The permissions granted to user 'Computer\User' are insufficient for performing this operation. (rsAccessDenied)

    Read the article

  • Website requests not reaching IIS?

    - by pete the pagan-gerbil
    To start off with a confession, I am not a server admin - just a developer tasked with getting to the root of a problem. Please be gentle! I have an intranet ASP.NET website running in IIS on a virtual machine. The website is not accessed very often (the last IIS log file was modified nearly six months ago). Both the IP address and Host header value are now failing to return the website, and the IIS log still doesn't show any more recent activity. The virtual machine was moved to a different physical location a few months ago, and the IP address for it has changed. Could this be what has broken access to the site? What else should I be checking to solve this? I don't have totally unrestricted access to the building's network settings, structures, etc. I would be grateful for any advice, even if I can't use it myself it'll improve my knowledge of what's going on behind the scenes!

    Read the article

  • Compressed filesystem inside a file in Linux [migrated]

    - by Doc
    I have a flash drive which is FAT32 formatted. I want to put a linux filesystem on the drive inside a file. I know I can do this by creating a file and formatting is with ext3 (or any other file system) and then mounting it with the -o loop option. What I would like is that the above filesystem be compressed. Essentially something like a read-write squashfs. Is there something that exits that I can use? Additional bonus if the file can be stored as sparse, i.e. the file re-sizes as data is written or deleted.

    Read the article

  • Windows 8.1 Insufficient storage available to create shadow copy

    - by Bob.at.SBS
    [Note: After I entered the problem statement, I found this question, which is apparently the same problem. Maybe one of us will get a good answer...] I have used the "Windows 7 File Recovery" tool under Windows 8 to create system image backups to an external USB hard drive. I built a new Windows 8.1 machine, and I want to create my first system image backup of that machine to the same USB hard drive. The "Windows 7 File Recovery" tool is gone in Windows 8.1, but wbAdmin is alive and well: wbAdmin start backup -backupTarget:\\?\Volume{2a2b...994f} -allCritical -quiet fails with this text displayed: wbadmin 1.0 - Backup command-line tool (C) Copyright 2013 Microsoft Corporation. All rights reserved. Retrieving volume information... This will back up (EFI System Partition),(C:),Recovery (300.00 MB) to \?\Volume {2a2b1255-3a86-11e3-be86-b8ca3a83994f}. The backup operation to F: is starting. Creating a shadow copy of the volumes specified for backup... Summary of the backup operation: The backup operation stopped before completing. The backup operation stopped before completing. Detailed error: ERROR - A Volume Shadow Copy Service operation error has occurred: (0x8004231f) Insufficient storage available to create either the shadow copy storage file or other shadow copy data. The EFI System Partition is 100 MB The Recovery Partition is 300 MB The C partition is 1.72 TB, NTFS, 218 GB used, 1.51 TB free The destination drive is 1.81 TB, NTFS, 678 GB used, 1.15 TB free I've fiddled with vssadmin resize shadowstorage, with no change in the error. vssadmin list shadowstorage displays: Shadow Copy Storage association For volume: (C:)\?\Volume{37a0...263}\ Shadow Copy Storage volume: (C:)\?\Volume{37a0...263}\ Used Shadow Copy Storage space: 2.39 GB (0%) Allocated Shadow Copy Storage space: 2.81 GB (0%) Maximum Shadow Copy Storage space: 531 GB (30%) Shadow Copy Storage association For volume: (F:)\?\Volume{2a2...94f}\ Shadow Copy Storage volume: (F:)\?\Volume{2a2...94f}\ Used Shadow Copy Storage space: 334 GB (17%) Allocated Shadow Copy Storage space: 337 GB (18%) Maximum Shadow Copy Storage space: UNBOUNDED (922154758%) (Yeah, the "percent calculation" for UNBOUNDED is seriously bogus.) I've run SFC /verifyonly and it seems happy. I've verified that the new `Volume Shadow Copy" service starts when I start the backup operation. Any suggestions?

    Read the article

  • Ubuntu 14.04 disk utility SMART self-test failed threshold not exceeded

    - by user2323470
    I'm using the "Disks" program in Ubuntu 14.04 (live DVD) to assess the health of a drive I suspect is failing. However, when I first opened the program, it showed that the overall health was OK and all assessments are OK as well. I then tried to run a short self-test, but now the overall assessment shows a red "SELF-TEST FAILED". In the details section it says "Last self-test failed (read)" and "threshold not exceeded". All individual assessments are still OK though!! What I don't understand is, does that mean that the test executed and determined that the drive is a goner, or does it mean that the test didn't actually execute properly?

    Read the article

  • How to setup a simple self-hosted dynamic DNS server

    - by Cerin
    I have a small internal network of physical machines running hypervisors, which in turn run several KVM Ubuntu virtual machines. How would I setup an internal dynamic DNS server so that when I run a script to create a new virtual machine, that VM could automatically register itself in the DNS server? Bind seems to be the standard DNS server for Linux, but it seems designed for a much more "static" DNS model. Dynamically updating this would require a complicated script that would have to SSH into the DNS server, edit configuration files, and then restart the server. This doesn't seem like a very elegant solution. Are there better options? I saw a similar question, although they're asking for a solution for a public setting on Amazon. My servers are entirely private, and I don't want to rely on an external VM host or Dynamic DNS provider.

    Read the article

< Previous Page | 353 354 355 356 357 358 359 360 361 362 363 364  | Next Page >