Search Results

Search found 6101 results on 245 pages for 'incremental backup'.

Page 27/245 | < Previous Page | 23 24 25 26 27 28 29 30 31 32 33 34  | Next Page >

  • Restoring file properties but not the complete files, from backup

    - by Jon
    While copying data from my old storage on a Linux computer to the new (linux-based) NAS, I accidentially failed with getting the properties (most important: the modify dates) along to the new location. I also continued to use/modify the files at the new location and hence, cannot just copy it all over again. What I would like to do is a diff between files in the old vs. the new storage, and for those being identical, restore the properties from Linux storage to the NAS storage files. Is there a clever way such as a script or a tool to do this? I could either run it on the Linux box or in worst case from a remote Windows computer. Grateful for any suggestions. /Jon

    Read the article

  • How restore qmail backup files

    - by Maysam
    We are using qmail as our mail application on a linux server. A few weeks ago our server crashed and we had everything installed from scratch and our users started to send & receive email again. The problem is they have lost their old emails. We have a back up of the whole qmail directory. But I don't know how to restore the old emails without losing the new ones. It's worth mentioning that I don't have any problem with restoring old sent mails. When I copy email files into .sent-mail/cur directory, I have them restored in sent box of users, but restoring files in /cur directory doesn't work for inbox emails and I can't get them restored.

    Read the article

  • Lazy linux backup system?

    - by Alex
    Ok, so I want to say Time Machine, but that's note exactly what I'm looking for. I want to set up a system that will regularly (hourly?) back up the /home directory of our machine. time Machine style things are naturally preferable since they save space by only saving the changes, but honestly, this is important enough that I can suffer some waste. Any ideas?

    Read the article

  • Using Windows Azure storage for backup

    - by Bruno
    I am currently looking at Windows Azure blobs as an option for backing up archive data. I want to be able to upload files from an external windows machine via the internet but I don't know enough about Windows Azure storage to make a decision. Some of the questions I have are How do I upload the files. Is there a client application, can I use robocopy? Would it be fast enough? i.e. Could I download or upload 1TB of data in a week? Is it secure? Hopefully someone smarter than me can help me :-)

    Read the article

  • How To Place a Drupal Site on to a server from backup

    - by CCG121
    I Backed up my site then totally re-did my server with a different Control Panel which created a different directory structure /var/www/vhosts/user/site.com/httpdocs I put the files into the httpdocs folder and put the database back correctly (I think) I can see the main page but clicking on any links gets me a Not Found Message I have tried running update.php and I cannot access /user/login either.

    Read the article

  • Schedules Folder Backup

    - by Junaid Saeed
    i have some folders in C drive on which i work on daily basis and the data in them is very critical.. so every night when i shutdown my PC i copy - paste - overwrite existing files these folders to a separate location... so that of the system crashed or something bad happens.. i will be able to easily format C and all i cannot move these folders from C drive because these folders include C:\wamp\www\ of WAMP server and such folders... is there a tool on which i can schedule that everyday at X time these folders will be backuped to 'Y' path

    Read the article

  • Backup with bash and rsync...

    - by Roger
    Is there a way to auto-rename an existing file on the receiver? For example: if filename already exists, it auto-rename filename to something like filename_001, filename_002 and so on.... So far all I have is this: $ rsync -rh --progress --stats --exclude '.thumb' \ --update --perms /origin /destination By the way, I know rsync has --ignore-existing to "skip updating files that exist on receiver", but I guess what I need would be something like --rename-existing.

    Read the article

  • Backup, Migrate or Clone Failing CentOS 4 (LVM)

    - by Hegelworm
    I've been running a BlueQuartz CentOS 4 system (Nuonce.net distro) for a few years now and although the hard drive (Deskstar) has always been a bit noisy, on a few recent occasions I've heard it having trouble spinning up. Basically, I want to clone this drive to a similar sized one (80 Gig). I've spent many hours reading upon dd, dd_rescue, rsync, clonezilla and LVM mirroring yet the sheer number of options and nightmarish accounts has left me frozen - unable to make an informed decision as to how to start. I've made a few attempts. dd failed after about 2 hours, as, although the drives appeared to be identical on the surface (ATA Seagate Barracudas, Thai not Chinese), the destination drive is slightly smaller. My most recent attempt involved using a Debian CD to format the new drive and then rsync-ing everything over and editing the new drive's grub and fstab to reflect the changes. No joy here either as I hadn't chosen LVM when partitioning the destination drive and it wouldn't boot. As you can probably tell, I'm out of my depth here and a panic-invoking mixture of caution and frustration has prompted me to sign up here. The server itself, although not strictly a production environment, has a very specific installation of Festival, LAME and ffMpeg and provides the back-end for a Text-to-Speech jQuery plugin that I've built over the last 2 years. I'm also planning to rebuild the whole TTS system on Debian as the existing CentOS system still has PHP4 etc. For now though, I'd really like to just shift everything over to a new drive. As this is my first post, please feel free to lay any house rules on me that I might've overlooked; I've been hovering around StackOverflow for a while now but have only just signed up. Many thanks. Update: Thanks for your responses so far - it's much appreciated and makes me feel a little more confident when I can double-check things here. I had the idea of doing a fresh install of CentOS (from the original disk) on the new drive so the partitions and LVM were all set up correctly (after disconnecting my source drive to prevent painful mistakes). I then booted into rescue mode from the same CD, and, to avoid a conflicting label, changed the /boot partition's label using e2label to /bootnew. I then changed the VolGroup name using lvm vgrename from VolGroup00 to VolGroup001. I could then boot with both drives in. After mounting the new drive (via its VolGroup001 alias) into /newhd, I rsync-ed over everything I could to the new drive, using -avr switches and backslashes. Like mentioned here. I then disconnected my original source drive again, booted from the liveCD again, changed back the boot partition label from /bootnew to /boot using e2label and then renamed the VolGroup back to VolGroup00. I then rebooted and it went through the familiar start-up routine only to not find a host of files in proc, usr, lib, var etc. The boot did complete but there were lots of red 'FAILS'. I could log in with my existing creds, but the network was kaput, I couldn't startX (desktop GUI) and there were also a few (a lot) of error messages pertaining to iptables. Back to square one. I naively thought I'd nailed it. Shall I just buy a bigger hard drive and attempt the dd route? I've read that this can mess with LVM setups and there's the added risk of working on two unmounted drives at once with a low-level tool. Thanks again.

    Read the article

  • Setting up logging for a remote backup script

    - by Brian Dainis
    So I wrote up a short script that I am planning to run via a cron job daily to package up my site files and send them to a remote location. I also plan to incorporate DB dumps, but I have not gotten that far yet. My issue today however is that Im am uncertain how to log the output of each command for errors, warnings, or other pertinent information the command may output. I would also like to install sometype of fail safe so if something goes horribly wrong the script will stop dead in its tracks and notify me via email or something. Ok the email thing is not as critical, but would be nice. Does anybody have any ideas for that? Here is what I have so far. By the way, both servers are CentOS 6.2 running standard LAMP. #!/bin/sh ################################# ### Set Vars ################################# THEDATE=`date +%m%d%y%H%M` ################################# ### Create Archives ################################# tar -cf /root/backups/files/server_BAK_${THEDATE}.tar -C / var/www/vhosts gzip /root/backups/files/server_BAK_${THEDATE}.tar ################################# ### Send Data to Remote Server ################################# scp /root/backups/files/server_BAK_${THEDATE}.tar.gz user@host:/home/bak1/ftp/backups/ ################################# ### Remove Data from this Server ################################# rm -rf /root/backups/files/server_BAK_${THEDATE}.tar.gz

    Read the article

  • stsadm farm backup exits with ffffffff

    - by overbyte
    I have a SP2007 farm that uses stsadm through Scheduled Tasks to run farm backups. It always worked fine, however it ran for a couple of seconds one day and just exited with code ffffffff. Looked at Event Viewer, the Sharepoint logs themselves and nothing unusual happened at the time this job ran. No files were created so an spbackup.log doesn't exist. Searched the net for batch files and STSADM return codes but the error message doesn't even exist. Any other recommended place to look for issues like this?

    Read the article

  • Best tool to backup your firefox shortcuts

    - by vaccano
    I have lost my shortcuts a few times (from hard drive crashes). Is there a good tool to back them up easily. (I would prefer to not have to remember to do it.) Backing them up to the internet would be a nice bonus, but it is not required for my needs.

    Read the article

  • How do I backup a git repo?

    - by acidzombie24
    I am planning to switch from SVN to git. With svn I just copy my repo folder when I want to back it up. However git doesn't have one so what do I do? Should I create a clone on a separate drive and update by pulling from my project? Then I can burn/archive this folder and it will have all the history? This is probably obvious but I want to make sure when it comes to backups. I still pretend there is a root repository.

    Read the article

  • Bootable backup for Windows (7) - Like Super Duper for Mac

    - by Dan F.
    Just got an SSD installed on my notebook and as people suggested I want have my bases covered in case it fails and I expect it to fail. Here is what I have in mind... keep a partition on the main drive (HDD) the same size as the SSD and keep a "clone" there, and in case the SSD fails... I take the SSD out and boot from the clone partition. From my understanding SuperDuper! does just that for MacOS, but I don't seem to find a something similar. I've found a lot of great tools out there that enables you to make bootable images (CloneZilla, DriveImage XML, Acronis® True Image™ to name a few), that is not what I'm looking for.

    Read the article

  • Are there free FTP serer backup repositories

    - by Saif Bechan
    I was wondering if there is a free service that provides a repository for your backups. These are backups of my server, there are usually about 200mb. I want to FTP the last 2 or 3. I am looking for a service that maybe provide some gigs of space and with FTP access. Looking at email providers such as gmail and hotmail that give you a couple of gigs of free space, this should also be possible or am i horribly wrong.

    Read the article

  • How can I capture a one-time full backup of a server using AMANDA?

    - by Daemon
    Suppose I have a preconfigured AMANDA server running automated network backups of directories specified in my disklists file. Normally, AMANDA will backup targets disks to /dumps/amanda. Is there any single command or method to perform a manual, one-time, full backup dump to another destination drive? I ask since I'm investigating the possibility of introducing rotating external hard-drive backups for offsite storage and I want to leverage our existing backup strategies wherever possible. Ideally, a full backup restore should be achievable from only any one of these offsite backup discs.

    Read the article

  • Can Time Capsule backup its own network share or can Time Machine be setup to do so?

    - by Sam Brightman
    Apple's Time Capsule can act as both a backup drive and a network shared drive at the same time. Can it backup its own data to incremental backups? I can't find this information anywhere. I've not used Time Machine so I don't know if it can be configured to backup shared drives (independently of whether Time Capsule has the option). I would like this to be a feasible way of freeing up some storage on an older MacBook Air by keeping large media such as photos archived on the drive like you would a USB external drive. But having the only copy accessible throughout the house and not backed up would make me nervous. I like the idea of having a good router, network storage and backup in one box, although obviously it's safer to backup the Capsule itself to USB occasionally too. Other solutions to this are of course welcome if this is not possible.

    Read the article

  • Is there an online storage service that works with Windows 7 Backup and Restore? [closed]

    - by user57813
    I love Windows 7 Backup and Restore.. However, I kind of find it inconvenient to use external hard drive, plug it in, take it out, put it away safely etc.. Is there an online storage service to which Windows Backup can save the backup files to? Can I make Windows Backup store the backed up data to an online storage service? I do not want to manually upload the backup files. I want it to save directly to the cloud. Doesn't matter if the service costs me some $$$. I am using Windows 7 Ultimate 32-bit edition.

    Read the article

  • Backup and Restore a SQLCE .sdf database

    - by Ruben Trancoso
    My application needs to backup and restore .sdf files. There is a single dataSet the the whole application and some bindngSource and table adapters on forms using this same dataset. Just for a sake of test I tryied to copy the .sdf in runtime for a backup folder and back to restore it and I got my application not finding the file like it was not there anymore. How should I manage connections to open and close the database since the dataSet do it automaticaly at begin and end of my application?

    Read the article

  • Online backup services

    - by Ilnur
    Could anybody recommend me free online service for everyday backup of my data. It contain files for about 5-10 GB. I'm looking for DropBox. It's very cool service. But I want to edit settings once, and then backup data automatically (my docs folder, music, photos and others). Any ideas?

    Read the article

  • MySQL Backup: Can I copying individual MyISAM table files to another server with different MySQL ver

    - by Komputer
    I means copying individual MyISAM table files is: (shut down mysqld and copy the .frm, .myd, and .myi files from one database folder to another) Question: (a) can I use this way to backup MySQL database folder from one server to another server with different MySQL version? (b) can this backup files moved to different OS? (example: debian to centos) thanks in advance.

    Read the article

  • MySQL database - backup problem

    - by Mike
    Hi I need to backup MySQL database and then deploy it on another MySQL server. The problem is, I need it backup without data , just script which creates database, tables, procedures, users, resets autoincrements etc. ... I tried MySQL administrator tool (Windows) and UNchecked "complete inserts check box", but it still created it ... Thanks in advance

    Read the article

  • One-click web site backup with asp.net

    - by HasanGursoy
    Hi, I want to backup my web site regularly but instead of backup my files via ftp program I want a one-click solution which will copy all my files and zip them and allow to download. Can this be done with asp.net at same web site or I have to write a .net app for this purpose?

    Read the article

  • Backup of folder + database - Python

    - by RadiantHex
    Hi there, I feel like this is quite delicate, I have various folders whith projects I would like to backup into a zip/tar file, but would like to avoid backing up files such as pyc files and temporary files. I also have a Postgres db I need to backup. Any tips for running this operation as a python script? Also, would there be anyway to stop the process from hogging resources in the process? Help would be very much appreciated.

    Read the article

  • Backup configuration of Hudson Master systems

    - by Praveen
    I need Backup configuration of Hudson Master systems Detailed information If a Hudson Master system goes down, we need to be able to bring it back up on a different VM as soon as possible (may be 3 hours). Therefore, I need to develop a way to backup the configuration and resetup a new VM reliabily. can any body let me know best way step by step. THANKS

    Read the article

< Previous Page | 23 24 25 26 27 28 29 30 31 32 33 34  | Next Page >