Search Results

Search found 7545 results on 302 pages for 'backup and restore'.

Page 79/302 | < Previous Page | 75 76 77 78 79 80 81 82 83 84 85 86  | Next Page >

  • Mountable online storage (no syncing)

    - by Sam
    I have a Linux VPS that I would like to turn into a media server. Like most cheap VPS's, it has a fairly small storage capacity. What I would like to do is attach the box to an online backup system such as SpiderOak or DropBox, where the files would reside and be directly accessible to either a webserver or media server software. Since the VPS hdd is small, I do not want the files to be synced to it. I would like a storage system that is online only. Ideally mountable like a network drive. Are there any services that suit my needs, or workarounds for services such as SpiderOak that do not require syncing?

    Read the article

  • Should I be running my scheduled backups as SYSTEM or as the our domain admin?

    - by MetalSearGolid
    I have a daily backup which is scheduled through the Task Scheduler. It failed with a strange error code last night, but I was able to search and find a blog post with how to avoid the error in the future. However, one of his recommendations was to run the backups as the Administrator user of the domain. Since all of the files being backed up are local to this system, should I continue to have the backups run as SYSTEM? Or is it actually better to run it as a different user? I have been running these backups for well over a year now and have only had a handful of failures, but ironically when it does fail, the error code means it was a permissions issue (or so I read, this code seems to be undocumented by Microsoft). Thanks in advance for any insight into this. Might as well post the error code here too, in case anyone would like to share their insight on this as well, but I rarely ever get this error, so I don't care too much about it: 4294967294

    Read the article

  • How to automaticaly mount luks-partition only when disk is plugged in

    - by Frederick Roth
    I have the following scenario: I want to automatically backup some data from my Laptop(Fedora Core 17) to a external encrypted(luks) hard disk. The disk can be opened by a key file, which lies on the also encrypted root partition of my laptop. The hard disk is attached to my docking station and therefore only "present" when I am at home (which is approximately 1/2 of the time the Laptop runs) I have everything set up the way I want it with one exception. I don't get a decent way to mount the hard disk automatically at boot if and only if it is present. If I add it to crypttab and fstab without noauto it tries to mount it at boot and takes a lot(!) of time and error messages when it is not present. If I add noauto, well it does not mount automatically ;) Is there a way to configure luks/crypttab to do the following: check whether the disk is present if yes: decrypt/mount if no: just don't

    Read the article

  • How to give a Linux user permission to create backups, but not permission to delete them?

    - by ChocoDeveloper
    I want to set up automated backups that are kept safe from myself (in case a virus pwns me). The problem is the "create" and "delete" permissions are the same thing: write permission. So what can I do about it? Is it possible to decouple the create/delete permissions? Another option could be to let the user "root" make the backups. The problem is my home directory is encrypted, and I don't want to backup everything. Any ideas? For the backups I'm using Deja Dup, which is installed by default in Fedora and Ubuntu.

    Read the article

  • Rolling Back Microsoft CRM during testing

    - by npeterson
    Process related question: Currently we have a multi-tenant installation of MS CRM 4.0 on three servers, Dev, Test, and Live. We are actively working on customizing one of the tenants, but the others are static. During user testing, we often find it necessary to 'start fresh' in one of the tenants. Is it better to try and delete out the changes from the tenant (created accounts, leads, etc), or just revert the database to a backup from before the testing started? Is there compelling reasons why bulk delete is not advisable for MSCRM or that reverting the database frequently could cause issue?

    Read the article

  • SQL Server 2005 SP3 Express Backups Incredibly Slow

    - by Adam Robinson
    I'm attempting to troubleshoot an issue with one of our customers who's using SQL Server 2005 SP3 Express to house their application data. The automatic backups that we perform when upgrading our application are timing out after 30 minutes, and I've been sitting and watching the backup take place in SSMS for about 20 minutes now and it's only gotten to 30%. The database is only slightly over 1GB, so I'm baffled as to what could be causing this sort of horrible performance. The machine is a 1.87GHz Xeon with 3GB of RAM running Windows Server 2003 R2. While that's hardly a powerful box, this seems ridiculous. Does anyone have any idea why these backups could be taking so long and, more importantly, how I can do something about it?

    Read the article

  • Using a script that uses Duplicity + S3 excluding large files

    - by Jason
    I'm trying to write an backup script that will exclude files over a certain size. If i run the script duplicity gives an error. However if I copy and paste the same command generated by the script everything works... Here is the script #!/bin/bash # Export some ENV variables so you don't have to type anything export AWS_ACCESS_KEY_ID="accesskey" export AWS_SECRET_ACCESS_KEY="secretaccesskey" export PASSPHRASE="password" SOURCE=/home/ DEST=s3+http://s3bucket GPG_KEY="gpgkey" # exclude files over 100MB exclude () { find /home/jason -size +100M \ | while read FILE; do echo -n " --exclude " echo -n \'**${FILE##/*/}\' | sed 's/\ /\\ /g' #Replace whitespace with "\ " done } echo "Using Command" echo "duplicity --encrypt-key=$GPG_KEY --sign-key=$GPG_KEY `exclude` $SOURCE $DEST" duplicity --encrypt-key=$GPG_KEY --sign-key=$GPG_KEY `exclude` $SOURCE $DEST # Reset the ENV variables. export AWS_ACCESS_KEY_ID= export AWS_SECRET_ACCESS_KEY= export PASSPHRASE= When the script is run I get the error; Command line error: Expected 2 args, got 6 Where am i going wrong??

    Read the article

  • What possible events could cause a MySQL database to revert to a previous state?

    - by justkevin
    A client of mine recently had a strange event with their MySQL database. Several days ago, one database suddenly "went back in time". All the data was in the state it was in several months ago. Even most of the .MYD and .MYI files had timestamps from November. Fortunately, the server is not in production yet, but we need to understand how it happened so it doesn't happen again. I'm not a MySQL guru, but I couldn't think of a scenario that could cause the database to rewind like that short of restoring from a backup. What could have happened here? Where should I look for clues? (Server is FreeBSD 6.4)

    Read the article

  • Best motherboard power supply combo for backblaze server

    - by jin14
    Building a backblaze server as described in this article. http://blog.backblaze.com/ So 45 hard drives in one box. I'm making it a MSDPM 2010 server so I actually don't even need raid cards in there as MSDPM will figure out how to use all of the hard drives on it's own. So need to know what motherboard, CPU, power supply I should get. Primary hard drive : SSD 128GB Storage : 45 1.5GB sata drives OS : windows 2008 Backup software : Microsoft System center Data protection server 2010 Need to know Which mother board to buy which will support 45 SATA hard drives. Don't need a raid card. Which power supply can power all 45 hard drives, 1 ssd drive, motherboard. Best set of equipment that meets my needs wins

    Read the article

  • Problem with script that excludes large files using Duplicity and Amazon S3

    - by Jason
    I'm trying to write an backup script that will exclude files over a certain size. If i run the script duplicity gives an error. However if i copy and paste the same command generated by the script everything works... Here is the script #!/bin/bash # Export some ENV variables so you don't have to type anything export AWS_ACCESS_KEY_ID="accesskey" export AWS_SECRET_ACCESS_KEY="secretaccesskey" export PASSPHRASE="password" SOURCE=/home/ DEST=s3+http://s3bucket GPG_KEY="gpgkey" # exclude files over 100MB exclude () { find /home/jason -size +100M \ | while read FILE; do echo -n " --exclude " echo -n \'**${FILE##/*/}\' | sed 's/\ /\\ /g' #Replace whitespace with "\ " done } echo "Using Command" echo "duplicity --encrypt-key=$GPG_KEY --sign-key=$GPG_KEY `exclude` $SOURCE $DEST" duplicity --encrypt-key=$GPG_KEY --sign-key=$GPG_KEY `exclude` $SOURCE $DEST # Reset the ENV variables. export AWS_ACCESS_KEY_ID= export AWS_SECRET_ACCESS_KEY= export PASSPHRASE= When the script is run I get the error; Command line error: Expected 2 args, got 6 Where am i going wrong??

    Read the article

  • Time Machine (archive) of pre-Leopard systems?

    - by benc
    I want to get off an older Mac OS X system permanently. It is an iBook G3, so that has two important characteristics: Power PC, not Intel based. Runs only Tiger, not Leopard. This means, as far as I can tell: Cannot run Time Machine directly. Here's the approach I have been contemplating: Mount the drive in Firewire mode. Back up the drive as a external drive to the Time Machine volume. Disconnect the drive (permanently). However, I'm concerned that this drive will eventually age out, when the Time Machine volume fills up, and the old-system-as-external-drive is gone. Would it be better to do a single backup with another utility, to shared disk?

    Read the article

  • How do I back up my Windows partition from an Ubuntu live CD?

    - by lalli
    My Windows partition (C:) is corrupt. I'm booting up from an Ubuntu live CD and trying to copy all the files from C: to my external drive, but the system expands all of the links, producing a projected copy size of 1.8TB (my external drive is just 1TB, and the data in c: is around 700MB). Then I looked at dd and other backup utilities. Anything I looked into, I couldn't figure out whether or not the image would be readable in Windows through any other app. Has anyone else tried to back up data from a corrupted Windows installation using Ubuntu?

    Read the article

  • data protector red tapes

    - by Caesar
    I am using HP Data Protector A.06.11 in my organization, with HP EML E-SeriesEML library, with 4 drives using LTO-4 tapes, and i am having some problems. Yesterday I put 5 new tapes in the robot and formatted them. At that time, the robot got just those empty 5 tapes with empty space. (all the rest of the tapes are red, or with protection) Today in the morning after the night (1 backup run at night), and 2 of the new tapes are red (the properties are): Writes : 2 Overwrites : 1 Errors : 9 I format one of them, and check for each drive if the tape become red, no one of the drives do it. In the main pool properties, in media condition got: Valid for : 36 (months) Maximum overwrites : 250

    Read the article

  • Is vSphere's Data Recover appliance 'production-ready'?

    - by Chopper3
    I have a smallish lab environment (16 x ESX4iU1 hosts and VC4U1) that I periodically want to backup. Normally in production we snap to secondary SAN boxes then have disk-based VTL backups via NetBackup which eventually migrate to off-site removable disks but this seems like an overkill for my own kit. I've spent a bit of time with vSphere's 'Data Recovery' appliance, it was easy enough to setup and I've not really ran into any issues with it but that doesn't mean I trust it fully. Have you had any experiences with it, positive or negative that would help me decide whether to trust it or pay Symantec for more licences? Thank you in advance.

    Read the article

  • Looking for fault-tolerant, multi-CD/DVD burning software

    - by MichaelKay
    A few years back, I saw an open-source application that created fault-tolerant, multi-CD backup sets. The thought was that if a few sectors went bad on one CD, the data could be reconstructed from others similar to a RAID 5 hard-drive setup. I did some searching and I could not find the application again. Does anyone know of a application that works in this manner. It needs to work with both CDs & DVDs. Linux or Windows will be OK.

    Read the article

  • Bootable and remote controllable imaging software?

    - by stefan.at.wpf
    I am looking for an imaging software, that one can boot from CD or usb stick (therefore without installing anything on the system!) and then control by some kind of remote desktop. So something like a combination of Acronis TrueImage's Bootable Media and Windows Remote Desktop. I don't even need to be able to store the backup on a network drive, as an external usb hdd is attached for these purposes. Background: A system that is normally controlled by SSH / command line and therefore has no keyboard/mouse connected that I could use to control Acronis TrueImage.

    Read the article

  • Cheap machine with Xen to have more than one personal dev servers with RAID option?

    - by AJ
    I do not have a lot of money so when I need a personal dev servers I decided I will use Xen on one machine to run various OS's. Although, I really like Dell Zino (http://www.dell.com/us/en/home/desktops/inspiron-zino-hd/pd.aspx?refid=inspiron-zino-hd&cs=19&s=dhs) as a very economical and compact PC which can be scaled to 8GB of RAM, I do not know how to have a RAID setup, for the backup. Is there an economical way to add a RAID setup to Dell Zino or I would have to invest in a proper box for that? Thanks for your help in advance. **Also, please let me know if this question is not meant for this forum.

    Read the article

  • Setting up Raid 1 Array for Home Server

    - by user1048116
    I'm not sure if this is even possible, but it's worth asking on here! Essentially I have a old machine at home (well, not old hardware wise, but I recently built a new gaming rig), which I decided to install a copy of W2008 R2 on and use as a file/backup server and media center'ish machine. As of now, it has a single drive partitioned into C and D, with D being the Data partition. I have happened to find an old 1TB SATA drive lying around at home, and was wondering if it's possible to setup a Raid 1 array in my rig within Windows without needing to lose everything on my first drive (or maybe even just mirror a specific partition, say the Data partition, as this is just what stores my photos etc). Maybe this isn't possible, but you never know :) Regards, T.C

    Read the article

  • Software that will burn DVDs with a SFV or Pararchive file for each disc?

    - by Matt
    I'd like to burn several thousand RAW files (.DNG) of around 10-30 MB each to DVD to backup my photo archive. I'm looking for software that can do this and include an SFV-type file on each disc burnt. These are my requirements: compression is optional, and probably not desirable due to the extra time involved files should be spread amongst the discs as self-contained units, i.e. I don't want to have to load files from more than one disc to be able to read the files on that disc, so that excludes WinRAR's spanning options I don't want to spend time writing ISO images first as this will be a task I'll need to repeat often as I add new images to my archive - the software should write to the DVDs for me as simply as possible I'd like the SFV/Pararchive/recovery record to be stored with the files on each disc, so it only references the files on that particular disk Thanks in advance!

    Read the article

  • Doing a mysql dump causes swapping issues

    - by DFischer
    I do a mysqldump manually every night. I just noticed that after it is done and I try to access the website it is very slow. After I take a look at the free -mh I notice that the server is now swapping when it otherwise wasn't before the mysqldump. What am I to do in this case? Just restart the server every time I backup? That doesn't seem very effective. My database file raw is 1.1gb after the dump.

    Read the article

  • How do I securely share / allow access to a drive?

    - by sleske
    To simplify backing up a laptop (Windows Vista), I'm planning on sharing its C: drive (with password protection) and using that to back it up from another computer. What are the security implications of this? If I share C: with a reasonable password, how big is the risk of compromise if the system is e.g. inadvertently used on a public WLAN or similar? Background: I'm planning to use [Areca Backup][1] to back up two systems (Windows XP and Vista). My current plan is to install Areca on the XP box, and share the Vista system's C: as a shared folder, so the XP system can read it. Then I can set up the drive as a network drive and have Areca read it like a local drive. Of course, if you can think of a more elegant way of doing this, I'm open to suggestions.

    Read the article

  • Rsync and Windows 7

    - by Nate
    Can someone give me any tips on setting up some sort of Rsync server/client on Windows 7 to run rsync between both my web hosting server, and a backup server that I have running Ubuntu? I've tried setting it up with this tutorial: http://www.youtube.com/watch?v=CvwdkZLNtnA Using copssh, and cwrsync. Ran into all sorts of troubles, including not being able to get cwrsync to run (it installs properly, but never starts up), and copssh not generating the keys at all. The guy was running Windows Server 2003, though, so I'm guessing the problems could just be because I'm running Windows 7. I've been trying to set it up with my Windows machine being the rsync server, and then Ubuntu and my webhosting VPS as the clients, but I realize it may be easier (and make more sense) to just setup the rsync server on Ubuntu, and then an rsync client on Windows 7? Can anyone point me in the right direction? I'm thinking of using this guide: http://www.gaztronics.net/rsync.php It seems a bit outdated, though.

    Read the article

  • Time Machine Server on Fedora

    - by Conner Stephen McCabe
    I was just wondering if anyone knew how to/or knew of a good guide to setting up mac Time Machine backup server on a fedora 17 machine. I recently ran into an issue with my HDD and lost all of my data, so I don't want it to ever happen again, as it was quite traumatising, haha! I just want to be able to back all of my data up on my fedora machine from my mac, at scheduled times when I connect to my home network. If anyone could help, it would be fantastic, I searched around on google, but all of the ones I found were for older distros. Thanks guys!

    Read the article

  • Change User Folder

    - by sazpaz
    So I had a backup and when restored it to my computer the User folder didn't actually change to use the backed-up folder, so now I have an owner folder (which is pretty much empty) and a my_username folder below C:\Users. How can I make my account use the my_username folder as the user folder? EDIT: I've tried creating a new account and then changing the name of my old folder to the name of the new account, but somehow Windows still knows it's not the 'real' user folder and creates a TEMP folder. In which registry is this configured?

    Read the article

  • DVD RW: Are they still relevant for backups?

    - by Harry
    Hello, With the availability of compact USB memory sticks with much, MUCH higher storage capacities is there still any use-case for taking periodic, incremental backups on DVD/RWs? The DVD/RW has an additional annoyance that you cannot drag and drop files to it as easily as you can on a USB memory stick. So, if I have a 4.7GB DVD/RW, I must re-burn the whole image every time I backup new stuff... with possibly rearranged file/folder structure. Secondly, why in this day and age you cannot install a file-system (like ext3 or FAT32) on a DVD/RW... and likewise on CD/RW's as you can on a USB memory stick? Many thanks, /HS

    Read the article

< Previous Page | 75 76 77 78 79 80 81 82 83 84 85 86  | Next Page >