Search Results

Search found 22449 results on 898 pages for 'complete pc backup'.

Page 126/898 | < Previous Page | 122 123 124 125 126 127 128 129 130 131 132 133  | Next Page >

  • Bash script to keep last x number of files and delete the rest

    - by Brady
    I have this bash script which nicely backs up my database on a cron schedule: #!/bin/sh PT_MYSQLDUMPPATH=/usr/bin PT_HOMEPATH=/home/philosop PT_TOOLPATH=$PT_HOMEPATH/philosophy-tools PT_MYSQLBACKUPPATH=$PT_TOOLPATH/mysql-backups PT_MYSQLUSER=********* PT_MYSQLPASSWORD="********" PT_MYSQLDATABASE=********* PT_BACKUPDATETIME=`date +%s` PT_BACKUPFILENAME=mysqlbackup_$PT_BACKUPDATETIME.sql.gz PT_FILESTOKEEP=14 $PT_MYSQLDUMPPATH/mysqldump -u$PT_MYSQLUSER -p$PT_MYSQLPASSWORD --opt $PT_MYSQLDATABASE | gzip -c > $PT_MYSQLBACKUPPATH/$PT_BACKUPFILENAME Problem with this is that it will keep dumping the backups in the folder and not clean up old files. This is where the variable PT_FILESTOKEEP comes in. Whatever number this is set to thats the amount of backups I want to keep. All backups are time stamped so by ordering them by name DESC will give you the latest first. Can anyone please help me with the rest of the BASH script to add the clean up of files? My knowledge of BASH is lacking and I'm unable to piece together the code to do the rest.

    Read the article

  • Is LiveDrive.com reliable?

    - by Marc
    I'm currently using DropBox (50GB account) which works fine, but at this moment I'm not impressed with its speed. I have a 60down/6up connection and only LiveDrive can use almost the full bandwidth of my connection. Dropbox often is very slow (avg. 100-500Kb/s compared to LD at 6MB/s). If I only look at the speed and storage costs then LD is much better, but I don't have enough experience with LD to be able to say something about reliability. Can anyone comment on this? Thx.

    Read the article

  • Help creating image from LVM

    - by jackhab
    I need to duplicate CentOS hard drive image for multiple stations. The HD has the following layout: Disk /dev/sdb: 250GB Sector size (logical/physical): 512B/512B Partition Table: msdos Number Start End Size Type File system Flags 1 32.3kB 107MB 107MB primary ext3 boot 2 107MB 250GB 250GB primary lvm I saved /dev/sdb1 to file with fsarchiver but for sdb2 I get: /fsarchiver savefs an2.fsa /dev/sdb2 oper_save.c#1006,filesystem_mount_partition(): can't detect and mount filesystem of partition [/dev/sdb2], cannot continue. removed an2.fsa Although fsarchiver probe simple correctly detects sdb2 as LVM2_member. Is fsarchiver correct tool for this job? What's wrong? I'm on Ubuntu 9.1 with fsarchiver 0.6.8 and lvm tools installed. Thanks.

    Read the article

  • Help creating image from LVM

    - by jackhab
    I need to duplicate CentOS hard drive image for multiple stations. The HD has the following layout: Disk /dev/sdb: 250GB Sector size (logical/physical): 512B/512B Partition Table: msdos Number Start End Size Type File system Flags 1 32.3kB 107MB 107MB primary ext3 boot 2 107MB 250GB 250GB primary lvm I saved /dev/sdb1 to file with fsarchiver but for sdb2 I get: /fsarchiver savefs an2.fsa /dev/sdb2 oper_save.c#1006,filesystem_mount_partition(): can't detect and mount filesystem of partition [/dev/sdb2], cannot continue. removed an2.fsa Although fsarchiver probe simple correctly detects sdb2 as LVM2_member. Is fsarchiver correct tool for this job? What's wrong? I'm on Ubuntu 9.1 with fsarchiver 0.6.8 and lvm tools installed. Thanks.

    Read the article

  • Restoring the exact state of a linux install to a different laptop with different sized drives and other hardware

    - by user259774
    I have an IBM running a Manjaro install that has already been used and settled into, with packages installed, browser profiles, etc, etc. The drive is 60gb, and it has a swap partition and an ext4 root partition. I need to move this profile to a Toshiba computer with a 320gb drive. How should I go about this? My inclination would be to shut down the toshiba, boot a live linux system, dd the whole 60gb drive to a file, boot the toshiba to a live system, then dd the file to its 320gb drive. Would this work? I know that it wouldn't with windows, but I believe this is an artificially imposed limitation from Microsoft. Is this correct, or is Linux similarly limited? If not, how could I go about this? Would clonezilla work, or would the hardware disparities prevent it from working?

    Read the article

  • How to copy directories using debugfs?

    - by STM
    The debugfs manpage gives the impression that the command 'rdump . .' will recursively copy all files found on the specified filesystem from the debugfs cwd to the native filesystem's cwd. Instead I seem to receive a syntax error, and no copy is initiated? These are the commands I run: cd /path/to/transfer/destination debugfs /dev/sda1 -R rdump . . My task is to copy the entire contents of a clean yet unmountable USB storage device to its host machine's HD. The host machine does not support the inode size used by the USB device's filesystem (256) and its software is not upgradeable, so my intention was to use debugfs to transfer the files. If anyone has any other suggestions for this task I'd be grateful.

    Read the article

  • How to make TimeMachine back up contents of any path or mounted volume

    - by Olfan
    I keep different types of data in different encrypted sparsebundle images (say, one for each client) which automatically mount upon login but can't be opened by anybody other than myself. So, after login I have a number of virtual volumes in /Volumes/ which keeps my client data both secure and organized. How do I include data inside these virtual Volumes in TimeMachine's backups, or data residing in any path on any partition/volume? I found a promising solution description at blog.eurocomp.info involving editing the com.apple.TimeMachine.plist but all I can get TimeMachine to do is backing up the sparsebundle files themselves. I want it to back up the files inside the mounted image, though - something like adding /Volumes/Client_abc/ to TimeMachine's search path. Please do not redirect my to this previous question as it doesn't solve the problem at all. Please also refrain from telling me why you think I should not want this answer as that will not solve anything either. Please lastly don't say "it can't be done" unless you can technically prove that claim.

    Read the article

  • Is it possible to rent a virtual instance of Windows/XP/Vista/7 in the cloud on pay as you go basis?

    - by Janusz
    I am a software developer and I need to test my software on different OS all the time. I am using free VPC images from IE6 compat kit from Microsoft but they have only limited options of Windows Versions (Win XP SP3 and some restricted version of Vista). I need a service where I could rent a Virtual OS in the cloud and connect via Remote Dektop. Use for couple of hours and then discard all the stuff I have been doing there. I tried using Amazon S3 but they only have Windows Server 2003 Anyone knows of such a service that would offer different XP/Vista/7 versions ?

    Read the article

  • How to copy all recovery points to another drive? - Norton Ghost

    - by chobo2
    Hi I have Norton Ghost and I have 2 drives dedicated to backing up my files. One is an external drive the other is internal. Now my internal drive has filled up with backups and I now want to copy all those backups to my external drive. However it seems to want me to do it one by one. Can I do it like a batch or some mass copy so it does one after another so I can say have my computer all night on? My plan was to fill up my internal drive - copy it to my external drive - fill up my external drive - copy it to my external drive. Once my external drive would be filled up I then would start deleting the oldest backups but that probably would give me like 6 months of backs that I can back through if I would ever need too.

    Read the article

  • Cloning hard drive -- data, operating system settings, everything

    - by Salman A
    I am using Windows XP. My hard drive (Seagate 160gig Barracuda) is about to fail. Its already developed bad sectors and it seems to get worse everyday. Data transfer mode is down to PIO mode 2, chkdsk runs every now and then, registry and important windows files get corrupted and I spend 30-60 minutes running chkdsk /f /r from the recovery console. I've got a replacement (Seagate 5000gig Barracuda) and now i want to transfer each and every thing on to the new drive. I don't want to go through windows and software installation, I spent ages getting all those software installed and configured on that hard drive. Need advice: whats the best way to transfer everything onto the new drive so that it behaves just like the old one. And are there any "gotchas".

    Read the article

  • Latitude XT2 Battery issue

    - by Multi_Mark
    I received the add-on battery yesterday and put it in side my new lattitude xt2 but it would not recognize the battery. the laptop simply flashed yellow and the battery itself when the white battery level indicator was pressed just flashed green i left it plugged in overnight and its still the same issue. it will not charge. i removed and re-inserted the battery several times to make sure the plug in was properly connected but still no luck. is there a trick to this? or some way i can make this work? or is it simply defective?

    Read the article

  • problems with WindowsImageBackup and write protected drives

    - by Ralph Shillington
    On Windows 7 I created a System image of my computer (C: and reserved partition) onto a USB drive. No problem. I then formatted the C: and installed the OS -- no problem Now I would like use the System Image and get back some of my documents etc. But I can't get access wo the WindowsImageBackup folder on the USB drive 1) Somehow the drive is write protected --- how did that happen? How do I unprotect that drive. 2) I can't access the WindowsImageBackup folder because I suspect the ACL is out of wack with my new SID. I would add my new SID to the ACL but I can't because the drive is write protected At the moment I'm completely disconnected from my files, which I thought (and still hope) are backed up. Understandably, panic is now setting in.

    Read the article

  • How do I download photos tagged of me from Facebook?

    - by Keith
    I want to be able to download (and back up) photos tagged of me in Facebook. I'm specifically not interested in my own photo albums - I uploaded them and therefore have them already in better quality than FB. What I want are the photos other have uploaded that have me in them. I have a couple of hundred of these now, and don't much fancy three hours of right-click save-as... There seem to be a couple of utilities that pop up with a quick search, but (call me paranoid) I'm wary about giving some random freeware app my login and password. Social safe looked promising, but as it doesn't support this feature at the moment it's kinda pointless. Can anyone recommend one that they've actually used? I'd consider an open source one - I'm a programmer and don't mind digging through to check that it doesn't do anything nasty.

    Read the article

  • HDD Carrier, like a soda carrier available at McDonalds?

    - by Jason Taylor
    We use external USB drives for backups, and they have to be stored offsite at the end of the week. Right now we have your standard external USB drive inside an enclosure. We were thinking about moving to a USB dock, and dock a bare HDD for backups, rather than having various sized and types of enclosures. If we were to do this, the drives need protection while being transported to/from the safety deposit box. Is there any kind of hard drive carrier that would let us slide two drives into it, and it would provide protection while the drives are carried around by non-technical people? I'm afraid such a product doesn't exist, but perhaps someone knows of something?

    Read the article

  • Is there an easy way to transfer PDF, Mobi, epub, ebooks to the iPad to read yet?

    - by Jian Lin
    Besides the 2 ways by using email or putting the PDF on a website (but it is a trouble to go to page 180 by repeated scrolling), is there an easy way to transfer the .PDF, .mobi, .epub to iPad to read, preferably to Kindle or iBooks on the iPad, so that we can add notes or hightlight or bookmark the pages? There seems to be way even on the O'Reilly page but it doesn't seem to work: http://oreilly.com/ebooks/mobi/ http://answers.oreilly.com/topic/1293-how-to-read-oreilly-ebooks-on-the-ipad-that-arent-from-the-ibooks-store/

    Read the article

  • Retrieving a virtual box vdh that has been deleted from within Virtual box itself

    - by WillNZ
    I had a MS Virtual machine that I imported to virtual box to see how it worked. I decided that in this case MS Virtual machine worked just as good so removed the virtual machine from Virtual Box. In the process I accidently removed/deleted the vhd file that MS Virtual machine uses. Can I get this back from somewhere within virtual box or the recycle bin or do I have to use a undelete utlilty?

    Read the article

  • Win XP Virtual Machine APP that won't START as a terminal services client

    - by Daniel
    I have a Win XP Virtual Machine APP that won't START as a terminal services client..... but if i disable terminal services inside the virtual machine, start the app then start terminal services... the APP runs fine (so long as terminal services are disabled when it starts.....!) My question is... is it possible to script things somehow so as to allow this app to be accessed directly from win 7 (with out having to manually run XP virtual machine)... I.E as I can not run it this way as it won't START with terminal services running.... I.e is it somehow possible to script things so the APP starts and terminal services is delayed then kicks in after a time delay.... (I know this may not be possible but I'm getting sick of all the mucking around...)

    Read the article

  • For large files compress first then transfer or rsync -z? which would be fastest?

    I have a ton of relativity small data files but they take up about 50 GB and I need them transferred to a different machine. I was trying to think of the most efficient way to do this. Thoughts I had were to gzip the whole thing then rsync it and decompress it, rely on rsync -z for compression, gzip then use rsync -z. I am not sure which would be most efficient since I am not sure how exactly rsync -z is implemented. Any ideas on which option would be the fastest?

    Read the article

  • SQL 2005 Log Shipping - Was working, now isnt!

    - by Jim
    Hello, I had log shipping working between two SQL 2005 server fine. I suspect that a job was added to the source server which backed up the transaction log to disk (nothing to do with the existing log shipping job). As I understand it, if you do this then log shipping will fail to work. Sure enough, it no longer works. I've deleted the job which had just been created. Log shipping still does not work. I've rebooted both servers and, again, Log shipping does not work. I'm at a loss now... all I get is the folloing error: The log shipping secondary database XXXXXXXXXX has restore threshold of 45 minutes and is out of sync. No restore was performed for 5882 minutes. Restored latency is 15 minutes. Check agent log and logshipping monitor information. Any help appreciated! Thanks in advance.

    Read the article

  • how to back up data from a machine that keeps hanging

    - by Amit Phatarphekar
    Hello - I have a storage server running opensolaris. But lately its been acting up - it hangs at random times due to some SCSI/ATA related error messages. I've tried to fix it without any progress, so I'm giving up now. The machine keeps hanging every 30 minutes or 1 hr ...sometimes after 4 hrs. Its very unpredictable. So I've decided to just reformat the storage server and start from scratch...maybe I'll just not use solaris and install something else, since the errors are related to solaris running on ATA HDD or something. Question - Before I reformat it, I want to back up some of the important data on it. Like it has a VM with 200 GB disk files, it has a whole bunch of ISOs stored on it etc etc. I'm using a simple scp to copy the files over to a different machine. My issue is that, because the machine hangs....sometimes my file copy is incomplete and I have to start all over again. Lets say I'm trying to copy a 200GB file which takes like 4 hrs....IF the machine hangs before the whole file i copied over...I have to recopy the file from scratch. Is there a solution to copy the files over such that if the machine hangs or network goes down..the copying can resume from where it left off? - like if 50 GB of a 200GB file was copied and machine hung....next time, it'll just continue to copy rest of the amount, instead of starting all over again. Thanks Amit

    Read the article

< Previous Page | 122 123 124 125 126 127 128 129 130 131 132 133  | Next Page >