Search Results

Search found 5810 results on 233 pages for 'offsite backup'.

Page 72/233 | < Previous Page | 68 69 70 71 72 73 74 75 76 77 78 79  | Next Page >

  • How do you apply development practices like version control, testing and continuous integration/deployment to system administration?

    - by arex1337
    Imagine you're going to manage a number of servers with a number of different services that's used by a number of people. Now say you want to reconfigure or replace some software on one of those servers. Obviously you don't want to work on servers that are in production. If this was a code change, as a developer, I would make the change on my local development machine, test it locally and commit the change to a version control system. The changes could then be deployed in a staging environment, tested further and finally deployed in a production environment. It would also be easy for me to roll back, if necessary. Generally, or specifically, how do you achieve this in system administration? (The first thing that comes to mind is to use virtual machines and put virtual machine images in version control, but I'm sure there is a lot of literature and clever solutions I'm not presently aware of.)

    Read the article

  • Problems using InfraRecorder to back up ISOs of certain CDs

    - by Voyagerfan5761
    I've gotten into the habit of backing up my CDs as ISO files, just in case the discs should be damanged, lost, or destroyed. Using InfraRecorder, the process is pretty painless. Unfortunately, I have run into at least two discs that don't back up. I get the error message: Can't read source disc. Retrying from sector 252270 Sometimes this will appear repeatedly. One of the discs is my retail copy of Star Trek: Armada II; the other is disc one of DOOM 3. Both discs run flawlessly when I put them in the drive and let Windows AutoPlay them. Armada II appears as two tracks (one data, one audio) in InfraRecorder, and the error happens at the approximate track boundary. DOOM 3's first disc, however, fails much sooner (around sector 990) and appears as one solid data track. Am I simply using the wrong tools for this job? InfraRecorder is a nice free tool that I can run from my flash drive and use for most tasks of this type, but it does seem to have trouble with certain things. Ideally I'd like to hear about any workarounds people have found for this issue, but if I must switch tools I'm open to it (preferably other free tools).

    Read the article

  • Forcing tar to create an empty archive

    - by snostorm
    I'm trying to use tar to tar files before transfer, so I can keep the entire file path rather than losing it along the way. However, when I try to tar an empty folder, it tells me that it is cowardly refusing to create an empty archive. I want to keep the empty folder on the other end, but don't want to put anything else into the archive to make it non-empty. Is there any way to do this?

    Read the article

  • How can I copy files from a DVD skipping corrupt files?

    - by ujjain
    I have so far tried the following software solutions without success. Windows Explorer Copy (default) TeraCopy Roadkil's Unstoppable Copier Unstoppable Copier stops copying as a whole, it keeps continuing the same file after 20 minutes, despite any settings. I have spent much time Googling, but the one program ("Unstoppable Copier") that people recommended, did not to the trick as it keeps choking on the same file despite a setting of "Fasteset Data Recovery", "Maximum Retries 0" and "Undamaged Files First". The back side of the DVD has some scratches. What options do I have? I have access to both Ubuntu and Windows 7.

    Read the article

  • Backing NTFS hard drive to a Linux distro

    - by Ricardo Ferreira
    Hey all, I want to move to Linux (Ubuntu more specifically) and also bought a new hard drive. What I'm not really sure about is if I put my old hard drive on an external case, then install Ubuntu on the new one and configure, etc... Can I easily access my old one (which is now an external hard drive)? Will I have no problems accessing my partitions? Thanks in advance.

    Read the article

  • Method for imaging a HDD? [closed]

    - by Sonny Ordell
    Possible Duplicate: Imaging new hard drive in Windows 7 laptop? I have to image my 320gb Laptop HDD before I send it in for repairs. The HDD is likely going to get replaced, and I would ideally like to be able to restore everything as I have it now without having to reinstall my OSes, programs and place all my files back again. I can make space on an external HDD I have, so am just looking for how I should go about this. Should I just use dd with a linux rescue cd? Or is there perhaps a more suitable program with its own rescue disk?

    Read the article

  • Do you know of an online back up & sync solution for Windows that actually works?

    - by Hermann
    I've been looking for some time for a program that backups your data to the cloud and is also able to sync 2 or more computers. By syncing I mean, if I change something on one computer that it should also apply the same change on another computer. I've tried two solutions so far which looked good initially but ultimately failed because they were buggy or simply don't work, or both. Syncplicity has messed up my file structure several times and they don't fix fix bugs that are literally over a year old. I then switched to SpiderOak which looked promising but ultimately had the same problems. Messed up my file structure several times, sometimes sync simply doesn't work, it's extremely inefficient and they keep pushing out releases were they say they fixed the sync bugs I reported but they did't. Does anyone of you know of an online sync solution that actually works?

    Read the article

  • Volume Shadow Copy Remotely?

    - by Wringley
    I'm currently running a Microsoft Server 2008 R2 box and want to utilize Volume Shadow Copy but my development machine doesn't have enough hard drive space for 30 days worth of copies, which is what I am aiming for and currently only have 5 days of backups. I have another Windows XP box and I was wondering if it was possible to remotely store the shadow copy images in my remote XP box. If it isn't possible with the built in Volume Shadow Copy, is there an open source alternative that I can utilize that has the same or similar functionality?

    Read the article

  • Tool for purging unneeded backups

    - by Dana the Sane
    I'm in the common situation where the one of the linux servers I use for storing backups is filling up. I'm wondering what tools are available for doing this. Ideally, what I would like is something that keeps nightlies for the previous month, weeklies for the 2nd to 5th preceding months and retains monthlies (well, every 3rd week) for an indefinite period. Everything that falls outside of that would be deleted after the backups are run. I could write a script to do this, but I feel like there must be a standard tool for this task.

    Read the article

  • i[Pod|Phone|Pad|*] backups in iTunes

    - by Maroloccio
    iTunes <- iPhone. At sync time, a back-up is performed. Which data is included, which data is not? i.e. are songs (potentially redundant) backed-up so that a computer ends up having both the source file on the filesystem and the copy within the device back-up? Is anything on the iPhone filesystem not backed up? (i.e. on a Mac using Time Machine, some files are excluded from the back-up even if not all of them can be recreated upon restore - I lost my postfix config this way..)

    Read the article

  • Where can I find ready to use windows scripts that used robocopy?

    - by Geo
    We are installing the Windows Resource Kit, and that installs RoboCopy. We want to have access to a few windows scripts that uses RoboCopy so we can start from those to build something else. Any ideas on where I can find a few samples? NOTE 1: A bit of information. Every time we try to copy D drive to E drive (new drive) we get an error that says: ERROR 32 (0x000000020) Copying File d:\pagefile.sys The process cannot access the file because it is being used by another process. Waiting 30 seconds. Just to help figure it out.

    Read the article

  • mirroring linux server to external usb harddrive

    - by DuPie
    My google-fu must be sucking. i havent been able to find a good solution for the following: numerous Linux server on commodity hardware Trying to do a recovery mirror copy to external harddrives External harddrives are smaller than source harddrives, but larger than data External drives are connected via usb2 (slow) Servers range from 20GB of data to 400GB of data Servers are remote, so hands on access is a pain need to copy boot files. empty external drives currently Basicly, looking for a way to do use a ghosting solution from INSIDE a running linux server to an external harddrive, without booting a cd etc. the rsync/cpio solutions i've looked at dont work great with grub/dev/proc etc. I understand that since the system isnt offline, it wont be a "mirror" image as files change, but thats ok. Are there any free/commercial products that would work?

    Read the article

  • Where Are Databases Located - MySQL File Location

    - by nicorellius
    I just installed a CRM application with a MySQL database. I thought I new the name of the database but I can't find it. Now I am trying to perform a mysqldump but I don't know the name of my database or where it's located. Most docs I read assume the admin knows where this database is located and thee name of it - I should know this, I know.

    Read the article

  • Download databasename.bak file

    - by Jordon
    I have downloaded databasename.bak file from my hosting company, when i tried to restore that DB file in SQL server 2008 it is keep on giving me following error. The media family on device 'C:\go4sharepoint_1384_8481.bak' is incorrectly formed. SQL Server cannot process this media family. RESTORE HEADERONLY is terminating abnormally. (Microsoft SQL Server, Error: 3241) According to this error and from following link http://www.sqlcoffee.com/Troubleshooting047.htm It is clear that either file i am downloading is corrupt or it is getting corrupted on the way? Any idea, why I am keep on receiving this error? I tried almost all ways but unable to fix this problem, please help me.

    Read the article

  • Limited access to Amazon S3 buckets

    - by Tomas Markauskas
    Is it possible to somehow limit the access to an Amazon S3 account. I don't really like the idea of distributing my secret access key to all of my applications, that want to access just a single bucket on my account. If someone gains access to one of the applications, I could loose all my data stored on S3. One way I was thinking to do it would be creating a second S3 account and give it access to just one bucket of the main account, but it's not really a great solution. Another nice thing for me would be to give the secondary account only write (but not modify/delete) and read access. That way I could upload backups or other files and be sure, that they won't get lost.

    Read the article

  • VSS Information

    - by MJ
    I'm currently taking care of backups for about 100 clients, and I'm really getting hammered with VSS issues. I've tried many different things, re-regsitering the VSS Dlls, reboots, patches, etc. What I want to know: Where can I find some detailed, technical information about VSS.

    Read the article

  • Mapping skydrive as network drive in macos

    - by vittore
    as you probably know, if you have windows live account you can use free skydrive 25 gb storage. Even more a lot of people know that if you go to your skydrive in browser and copy cid query parameter value (https://...live.com/...&cid=xxxxxxxx ) you will be able to map skydrive as network drive in windows using this network pass \[cid].docs.live.net[cid]\ I do now that if you have network share like \server\folder i can map it in macos too, as smb://server/folder. however it is doesn't seem to be a case with skydrive when i try to map it as smb://[cid].docs.live.net/[cid] finder tells it can't connect. Anyone know how to map it ?

    Read the article

  • Linux: Alternative to rsync? (ie, scp with resume)

    - by Joernsn
    I've been using rsync to automatically send files from one box to another, which is great compared to scp, since it supports resuming. However, when resuming a very large file (10gb) rsync has to read both files and compare them, which is very slow. I don't need fancy error handling, just "scp with resume", so here's my question: Is there an alternative to rsync/scp, that supports resuming without having to read both source and destination files? I've read the manuals without finding anything I can use, please let me know if I've missed something. This is the rsync line I've been using: rsync -av --partial --progress --inplace SRC DST

    Read the article

  • Time machine folders won't restore recursively

    - by Brian Postow
    I have a Snow Leopard system with Time Machine. Every so often I need to look at an old folder, so I go in and try to restore it, usually to a different location. What I end up with is an empty folder of the appropriate name. none of the files, nothing. It doesn't give any error messages. The files ARE there, because I can see them, also, if I go in via the mount point, and trace down through the file system, I can do a filecopy of the folder, and everything turns out fine. However, that seems like a bad idea. so, I'd like to know why he "right" way doesn't work... I believe that it used to work. It's also possible that it only doesn't work for certain folders. I haven't tested extensively.

    Read the article

  • Should I use "Raid 5 + spare" or "Raid 6"?

    - by Trevor Boyd Smith
    What is "Raid 5 + Spare" (excerpt from User Manual, Sect 4.17.2, P.54): RAID5+Spare: RAID 5+Spare is a RAID 5 array in which one disk is used as spare to rebuild the system as soon as a disk fails (Fig. 79). At least four disks are required. If one physical disk fails, the data remains available because it is read from the parity blocks. Data from a failed disk is rebuilt onto the hot spare disk. When a failed disk is replaced, the replacement becomes the new hot spare. No data is lost in the case of a single disk failure, but if a second disk fails before the system can rebuild data to the hot spare, all data in the array will be lost. What is "Raid 6" (excerpt from User Manual, Sect 4.17.2, P.54): RAID6: In RAID 6, data is striped across all disks (minimum of four) and a two parity blocks for each data block (p and q in Fig. 80) is written on the same stripe. If one physical disk fails, the data from the failed disk can be rebuilt onto a replacement disk. This Raid mode can support up to two disk failures with no data loss. RAID 6 provides for faster rebuilding of data from a failed disk. Both "Raid 5 + spare" and "Raid 6" are SO similar ... I can't tell the difference. When would "Raid 5 + Spare" be optimal? And when would "Raid 6" be optimal"? The manual dumbs down the different raid with 5 star ratings. "Raid 5 + Spare" only gets 4 stars but "Raid 6" gets 5 stars. If I were to blindly trust the manual I would conclude that "Raid 6" is always better. Is "Raid 6" always better?

    Read the article

  • What is the fastest way to clone an INNODB table within the same server?

    - by Vic
    Our development server is a replication slave of our production server. We have a script that developers use if they want to run their applications/bug fixes against fresh data. That script looks like this: dbs=( analytics auth logs users ) server=localhost conn="-h ${server} -u ${username} --password=${password}" # Stop the replication client so we don't encounter weird data. echo "STOP SLAVE" | mysql ${conn} # Bunch of bulk insert optimizations echo "SET autocommit=0" | mysql ${conn} echo "SET unique_checks=0" | mysql ${conn} echo "SET foreign_key_checks=0" | mysql ${conn} # Restore all databases and tables. for sourcedb in ${dbs[*]} do destdb=${prefix}${sourcedb} echo "Dropping database ${destdb}..." echo "DROP DATABASE IF EXISTS ${destdb}" | mysql ${conn} echo "CREATE DATABASE ${destdb}" | mysql ${conn} # First, all the tables. for table in `echo "SHOW FULL TABLES WHERE Table_type <> 'VIEW'" | mysql $conn $sourcedb | tail -n +2`; do if [[ "${table}" != 'BASE' && "${table}" != 'TABLE' && "${table}" != 'VIEW' ]] ; then createTable=`echo "SHOW CREATE TABLE ${table}"|mysql -B -r $conn $sourcedb|tail -n +2|cut -f 2-` echo "Restoring ${destdb}/${table}..." echo "$createTable ;" | mysql $conn $destdb insertData="INSERT INTO ${destdb}.${table} SELECT * FROM ${sourcedb}.${table}" echo "$insertData" | mysql $conn $destdb fi fi done done echo "SET foreign_key_checks=1" | mysql ${conn} echo "SET unique_checks=1" | mysql ${conn} echo "COMMIT" | mysql ${conn} # Restart the replication client echo "START SLAVE" | mysql ${conn} All of these operations are, as I mentioned, within the same server. Is there a faster way to clone the tables I'm not seeing? They're all INNODB tables. Thanks!

    Read the article

< Previous Page | 68 69 70 71 72 73 74 75 76 77 78 79  | Next Page >