Search Results

Search found 14013 results on 561 pages for 'remote backup'.

Page 114/561 | < Previous Page | 110 111 112 113 114 115 116 117 118 119 120 121  | Next Page >

  • How can I use rsync with a FAT file system?

    - by Kim
    I would like to write a simple backup script that saves some data to a FAT drive. Should I reformat the drive and use a better file system or is it possible to use rsync with FAT? If so, what problems might I run into? Would performance be a lot worse? EDIT: This is on linux, didn't even know there was a rsync for windows. The sources are various file systems (it's a mess), and the destination is currently formatted with FAT32. Thank you for your answers, I'll probably go for a reformat, since I'm not completely sure about the file sizes we'll have.

    Read the article

  • How to repair a Veritas tape that has been overwritten a bit?

    - by Ismo Utriainen
    I meant to restore some files, but I forgot that there was a monday backup job just waiting for tape loading. So veritas 10d started to write over my tape and that valuable data is now gone. The original data size was about 40 GB and that accidentally started job wrote about 30 MB to the begin of tape. What are my possibilities to recover some data from that tape? Update: Inventory and catalog doesn't help, media settings are overwrite, not append. It is a DLT drive.

    Read the article

  • One way sync with either Dropbox or Google drive on linux and windows

    - by sup
    I could not google this one. I would like to use Dropbox or Google drive only as backup, so I would like to ensure I only upload to those services and never download (unless done manually via the web interface). There would be several of us uploading stuff to one account, so we would need to ensure we don't accidentaly delete something on another guy's machine by doing some changes on our local machine. Is there a simple way to do this automatically - i.e. by running a daemon that will upload eveyuthing in a given folder but never download anything? I am on linux, so that complicates things a bit, but I am also interested in windows solutions.

    Read the article

  • Any freeware/ideas for getting Windows 2008's backups to dump to tape after backing up to disk?

    - by TheCleaner
    I have a Windows 2008 R2 server that is being backed up to an external ISCSI drive nightly. The problem is, we'd like to use our Tape Drive (VXA 320) that Windows sees just fine to take those backups in the "WindowsImageBackup" folder and dump them to tape once a month so that we can at least have something offsite. I really don't want to go through the hassle of licensing BackupExec or similar if possible. All I'm really after is some kind of copy utility that can copy the "WindowsImageBackup" folder over to the tape drive. Ideas? P.S. If by doing this it wouldn't matter for a restore regardless, then let me know, but I would assume I could copy the folder back over to the server and then have Windows Backup find it again.

    Read the article

  • Can Windows Home Server be used on an active directory domain?

    - by Parvenu74
    The situation: an Active Directory network with a few dozen machines. Most of the machines have the same vanilla image applied to them so if there was a hard drive failure getting the machine back up to the standard network image would be quick and easy. However, there are a handful of (eight) machines which have rather unique setups (accounting, developers, the "artist" with CS4 and such). For these machines we would like to use Windows Home Server since the backups are automatic and recovery from a machine failure is quite painless. The question though is whether or not WHS can be used on an A/D network. If not, what "set it and forget it" backup/imaging product is recommended for this scenario?

    Read the article

  • Whats the best way to deal with backups for my php/mysql application

    - by spirytus
    I'm creating php application for my client and now thinking what would be the best way to do backups, automatically if possible? I don't have much experience in this area and in case something goes wrong, or if I need to migrate, I would like to have fast way of getting it all back online. I understand "something goes wrong" is a very wide term, but lets say that someone hacks my site and wipes out database and all the files. My app. is written in php/mysql and I got access to cpanel (hosted with justhost.com if that makes any difference :). I used Joomla and it has JoomlaPack that does complete backup almost automatically and in case site fails, its easy to revert, or migrate if necessary. Is there anything like that for my configuration that would make reverting/migration, easy?

    Read the article

  • Drobo Pro or SAN for backups

    - by PHLiGHT
    I have multiple offices and was thinking of using two Drobo Pros as a storage location for my backups, with one at each site being synced. It seems like on the surface its a cheap alternative to a SAN. Because it will just be used for storing backups I don't think that performance is as big of an issue. Related question. How do you sync your backups from your main location to a remote site. I am using Backup Exec and wondered if I should use a feature in the program or use something like Delta Copy.

    Read the article

  • Distributed filesystem for automated offline data mirroring

    - by Petr Pudlák
    I'd like to achieve the following setup: Every time I connect my laptop to a local network, my partition gets automatically mirrored to a partition on my local server. I only want to mirror what has changed from the last time. (I understand that it is not a proper backup solution since there is no history of the changes, it'd be more like a non-persistent network RAID.) Is there a distributed file system that allows such a setup? I've done some searching and it seems to me that most distributed file-systems are focused on data availability and distribution, not duplicating them. I'd be thankful for suggestions. Edit: Sorry, I forgot to mention: I'm using Linux.

    Read the article

  • Image an entire computer

    - by DalexL
    I want to start doing a computer repair. I know that some issues may arise (what if I do something wrong and ruin the entire computer?). I'm not saying I'm not confident in my work, but accidents do happen sometimes. Now, the best solution this problem would be to simply just backup their entire computer and restore it if anything goes wrong. If I have an external HD, are there any available methods to create an entire image for the computer so no matter what I do, I can just reload it's last state?

    Read the article

  • How does dropbox version/upload large files?

    - by barfoon
    Hey everyone, I have a free dropbox account (2GB), and I was wondering how the versioning of large files works. I have a full backup of all my webfiles that sites @ just over 1GB. After the initial upload of 1GB, everytime it syncs will dropbox figure out the delta of the file, or will it have to upload the entire thing again to version it? It would be cool to always have an up to date version of a large file, but I dont want to kill my bandwidth uploading 1GB everytime. Is this possible? Thanks,

    Read the article

  • How many hours of use before I need to clean a tape drive?

    - by codeape
    I do backups to a HP Ultrium 2 tape drive (HP StorageWorks Ultrium 448). The drive has a 'Clean' LED that supposedly will light up or blink when the drive needs to be cleaned. The drive has been in use since october 2005, and still the 'Clean' light has never been lit. The drive statistics are: Total hours in use: 1603 Total bytes written: 19.7 TB Total bytes read: 19.3 TB My question is: How many hours of use can I expect before I need to clean the drive? Edit: I have not encountered any errors using the drive. I do restore tests every two months, and every backup is verified. Edit 2: The user manual says: "HP StorageWorks Ultrium tape drives do not require regular cleaning. An Ultrium universal cleaning cartridge should only be used when the orange Clean LED is flashing." Update: It is now May 2010 (4.5 years of use), and the LED is still off, I have not cleaned, backups verify and regular restore tests are done.

    Read the article

  • Should you archive documents before backing up to the cloud?

    - by gabbsmo
    I'm planning to add a cloud storage to my personal backup strategy. But now I wonder if it really is worth the trouble of compressing my documents and photos. The Open XML-format already have zip-compression and JPEG is a lossy image format. So there really isn't much benefit in compressing. 20Mb of documents become about 17Mb at the ULTRA preset of 7-zip. One benefit I can imagine is that you can shorten upload time by archiving the folders since it minimizes the number of requests that is needed to be sent to the server at upload and download. So what are your thoughts and experience in this issue? Should you archive your documents before backing them up to the cloud?

    Read the article

  • rsync (or robocopy) from multiple computers - what happens?

    - by TheCleaner
    If a Linux server has two different rsync jobs nightly for the same folder to two different destinations, do both destinations end up with the same end set of files? Or does the first job run, and set something on the source folder/files that would cause the 2nd rsync job to not realize the daily changes/updates to the source? Same for a Windows environment using something like robocopy, or even a "differential" backup using BUE or similar. Does each "sync" compare the destination to the source and update the destination regardless of if it is synced multiple times to different destinations?

    Read the article

  • Is there a Utility that will scan selected locations and return all files older than a certain date?

    - by CT
    Can anyone recommend a utility that can scan specified directory locations (network shares specifically) and return all files older than a certain date? I am looking to implement a data retention policy at my workplace. As our amount of data grows it puts a large strain on our backup routines. I would like to move old data to some sort of archival system. Extra points for the ability to move queried old files to another location for archival and the ability to create schedules for when this occurs. Many thanks. EDIT: Windows Shop. Mostly Windows 2003 Servers.

    Read the article

  • Moving Windows 7 OEM from one hard disk to another but retaining the rest of the hardware

    - by Dane
    I think this should be very easy but it doesn't seem like there's a whole lot of info out there. I've personally never done this exact maneuver because I've never needed to retain the OEM licence before. I have a laptop and I wish to simply swap out the hard disk with a bigger one but keep the Windows installation and files intact. Do I use Windows Backup and Restore and burn DVDs or do I just Acronis or something similar. Or is there an even easier way of doing this involving a third party desktop? Any help much appreciated.

    Read the article

  • Cloning Fresh Windows 7 with external HDD, DVD or CDs?

    - by hhh
    I want to create Windows installation disk (not necessarily CD) from my Windows 7 Pro laptop, not sure what it actually means here. The material is about 50-60GB, requiring about 12 pcs of 5GB disk (not going to happen, too much work to use them later). I have the Windows serial on the laptop bottom and this is so-called firm -laptop (no idea what it actually mean, better warranty and some instant support thing apparently). Now how to do the clone with external material such as DVDs? how can I create mock-windows-installation medium or some real windows-installation medium? I am now not sure whether Windows offers W7 -installations medium online, well my W -peer mentioned some MSDAA -something. Perhaps related Cloning Fresh Windows 7 -fsed HDD to Linux Server because having no external HDD or disks for the Backup

    Read the article

  • Creating image of NTFS partition in Ubuntu

    - by Pappai
    Hi, I have a system that has to be formatted often because of the specific use of it (Use it in internet cafe). I have to install the drivers and apps each time. That is a time consuming and cumbersom task for me. I want to install windows, drivers, apps etc. once and create a backup of the entire C:\ drive, and keep it in a linux partition so that I can restore the OS with all the apps & drivers ready to go! I have ubuntu live CD with me and I have created a linux partition (ext4) in the HDD. My question is: How can I create an image of the C:\ drive (ntfs disk) in Ubuntu and store it in the linux partition?

    Read the article

  • rsync generates very much traffic

    - by user109459
    I use rsync for backing up one of my servers with 4GB of files. When I now try to transfer these files the traffic for the files isn't the estimated 4GB. It is a lot higher. It's about 60GB. I also checked the traffic on my server, backup server and router and all three say that there was a traffic of 60GB. But at the end rsync says that it only has transfered 4GB. Another problem is that I can't debugg it because the problem occures randomly.

    Read the article

  • Nightly backups (and maybe other tasks) causing server alerts

    - by J. Pablo Fernández
    I have two independent alert notification systems for my servers. The server is a virtual machine on Linode and one of the alerts comes from Linode. The other monitoring system we use is New Relic. They are both watching out for IO utilization. Every night I get alerts from both of them as the server is using too much IO. I run quite a few tasks in the middle of the night but the one I confirmed that can cause IO-warnings is running the backups. The backup is done by s3cmd sync. I tried ionice but it still generates the warnings. Getting warnings every night reduces the efficacy of warnings when they happen for real. For Linode I could raise the level at which a warning is issued, but it might mean making the whole thing useless as the level is too high. What would be the proper solution for this?

    Read the article

  • Back up Linux environment

    - by joesavage
    I'm currently in the process of installing a bunch of stuff I need and getting setup my Linode that I recently purchased. Being a Linux newbie, I'm doing pretty well - but one small mistake can screw everything up for me. I've currently got apache2 and some other things successfully installed and setup how I want them, and DO NOT want this to be ruined by some newbie mistake. What is the easiest way to backup the actual environment itself so that I can restore the backed up environment (with apache2 and things fully working) if I mess up?

    Read the article

  • Should the virtualization host be allowed to run any service?

    - by Giordano
    I recently setup a virtualization server for the small company I'm running. This server runs few virtual machines that are used for development, testing, etc... My business partner works from a remote location, thus I also installed a vpn server on the virtualization host to make it possible for him to safely reach the company services. Moreover, again on the virtualization host, I installed bacula to perform the backup of the data. Is it advisable/good practice to do so or should I create one more virtual machine to do backups and VPN? Is it a bad idea to run these services on the host itself? If yes, why? Thanks in advance!

    Read the article

  • tar gzip slowing down server

    - by Josir
    I have a backup script that: compress some files generate md5 copy the compressed file to another server. the other server finishes comparing MD5 (to find copy errors). Here it's the core script: nice -n 15 tar -czvf $BKP $PATH_BKP/*.* \ | xargs -I '{}' sh -c "test -f '{}' && md5sum '{}'" \ | tee $MD5 scp -l 80000 $BKP $SCP_BKP scp $MD5 $SCP_BKP This routine got CPU at 90% at gzip routine, slowing down the production server. I tried to add a nice -n 15 but server still hangs. I've already read 1 but the conversation didn't help me. What is the best approach to solve this issue ? I am open to new architectures/solutions :)

    Read the article

  • Settings what-opens-what once and for all (Backing up File Associations)

    - by ldigas
    Every time I switch machines (as in, get a new one, or reinstall an OS or something like that) my precious file associations get lost. And the next six months pass slowly until I again set them up right. Is there a program that allows me to: Set all the extensions I would like to open with let's say, Vim, without setting each one of them individually. Something of a kind: Vim opens: .... list of extensions ... and/or A program which lets me backup my current settings, and when I copy those to a new machine it lets me just modify the paths where I putted the applications in question, and it does the rest (again, associates that program with all the extensions it opened before).

    Read the article

  • Why is Ubuntu One slow to sync in 11.10, either backup or any sub-folder contents?

    - by pst007x
    I have been trying to sync my documents folder of 1.4GB, it still hasn't worked and it has been syncing for a month. The top level syncs, files and folders in the Document folders, but contents of sub-folders just hang. (Gave up and stopped syncing this folder) However,I have tried using the backup facility in 11.10, to backup to Ubuntu One.... I upgraded my HDD space in Ubuntu One. It has been going now for 24hours-ish and only backed up what looks like a couple of percent. (By the way what an excellent idea to backup to Ubuntu One, if only we could get it to actually work! :-o) The odd thing is I can sync to drop box within hours, rather than months. This is bad, and has been an issue since Ubuntu One's release. I have reported this problem and there were promises in later releases this would be fixed, but it hasn't. Canonical cannot help either... I posted on several blogs, a lot of people have the same problem but no fixes. So do I use dropbox or another service, until it is sorted, as Ubuntu does not seem to see this as an issue, I think a fix will be a long time in coming. (However,I love the potential of Ubuntu One and the integration with the OS) Yes my internet speeds are fine, etc... :-) No firewall (sudo ufw status: STATUS: INACTIVE), No Proxy, etc NB: I have raised this as a separate question to others posted here, because my question relates to Ubuntu 11.10, though I have commented elsewhere for help. Plus my question also relates to deja-dup backup to Ubuntu One. Thanks

    Read the article

  • LogonUser using LOGON32_LOGON_NEW_CREDENTIALS works against remote untrusted domain machine

    - by Jiho Han
    So between the two machines, there is no trust - they are in different domains. I've successfully connected to the remote machine using LogonUser API using logon type, *LOGON32_LOGON_NEW_CREDENTIALS*. I am able to retrieve the content of a directory using the UNC share, and create a file stream to "download" the file. So far so good. The only issue is that it seems, LogonUser fails unless there is an already open session. Let me clarify that. I found that the ASP.NET MVC page was not working this morning, specifically the page that retrieves the file list from this remote machine using LogonUser. I look at the log and I see in the stacktrace, *System.IO.__Error.WinIOError* above Directory.GetFiles call. I then remoted into the web server and tried to open the remote folder in the explorer using the same login/password used by the web site. It went through and I could see the files. I opened up the command prompt, type in net use, and I see that there is an open connection to the remote machine. Then I went back to the page and suddenly the page is working again. So, at this point, I am not exactly sure if the LogonUser is working as expected or not. If the call requires that a network connection opened first by other means, then this is certainly not satisfactory. Does anyone know what may be happening or suggest a workaround?

    Read the article

< Previous Page | 110 111 112 113 114 115 116 117 118 119 120 121  | Next Page >