Search Results

Search found 1833 results on 74 pages for 'floppy disks'.

Page 43/74 | < Previous Page | 39 40 41 42 43 44 45 46 47 48 49 50  | Next Page >

  • Windows data backups with alternating removable drives?

    - by luke
    0 vote down star I have a removable disk drive (RD1000 from Dell), and I am looking for backup software that will allow me to backup every night, and every morning switch to the alternate disk. There is only one directory structure to back up, what I want is two copies, one which I will take home with me every night, one which will be backing up every night, and when I get in in the morning, I will switch them. So for instance I have disk "a" and disk "b". On Monday night I want to go home with disk "a" and leave disk "b" in the drive, so that a scheduled back up will be written to it. On Tuesday morning I will come in and swap disks, and I will take "b" home that night, leaving "a" for the backup. And so on for the remainder of the week. FOSS software preferred, Freeware acceptable, paid software as a last-ditch effort. Oh, btw I'm stuck with Windows 2000 Thanks in advance.

    Read the article

  • smartctl or hddtemp for xvda [on hold]

    - by HST
    I'm trying to check the state of the drives on a remote server running Debian wheezy. I'm using a software RAID10 on top of, I guess, xen, since the entries in /dev are /dev/xvda and /dev/xvdb But it I try smartctl -a /dev/xvda I get /dev/xvda: Unable to detect device type Smartctl: please specify device type with the -d option. I've tried various device type guesses, none work Similar problem with hddtemp, which reports ERROR: /dev/xvda: can't determine bus type (or this bus type is unknown) I've searched the smartmontools documentation, but can't find any discussion of virtual disks. . . How do I get behind the virtualisation to something smart tools or hddtemp can work with?

    Read the article

  • NTFS disk mounted as fuseblk in ubuntu 12.10 is very slow and a lot of errors when rsync. Is that not a rare thing?

    - by Pablo Marin-Garcia
    I am having problems with a NTFS disk mounted as a fuseblk in my ubuntu 12.10 through external usb3. When I did a 1.1TB backup with rsync the speed was 1-2MB/s (wiht a ext4 disk speed was 70 MB/s before and after trying the NTFS disk). Also after one hour errors started to appear: rsync: write failed on "xxx": No such file or directory recv_files: "yyy" is a directory #but this file is a FILE not a dir ??!! .... As this is the first time I have mounted the NTFS in linux for heavy usage (the data would be used in windows afterwards), I would like to know if this kind of thinks are common o was only that something became unstable in my system and a simply restart would probably have solved it. This leads me to the these questions: Can I trust fuse for manage NTFS disks? Or is a problem of the NTFS tools in linux not yet totally stables for writing? Do people is still suffering from low performance with fuse-NTFS vs ext4 (in the past I have read about people complaining about this)?

    Read the article

  • Unmounting a zfs pool while it is shared with sharenfs

    - by Ted W.
    I have a Solaris (open indiana) system which is getting poor disk write performance. In order to enable ZIL in this version of zfs I need to add a line to /etc/system. This will not take affect until I've unmounted and remounted the zpool. The trick is that this spool is shared via nfs to about 200 other servers to host users' home directories. I can guarantee that no users will be accessing the disks during this period of maintenance but I would like to avoid having to issue an unmount for 200 systems in order to unmount the disk on the Solaris box. My question is, with sharenfs, is it necessary to have all systems disconnected before unmounting the filesystem on the host? If it's possible, how do you go about it? I've tried unmounting already, the normal way, and it reports the disk is busy. There is no lsof in Solaris and pfiles (I think that's what it was) does not show anything obviously using the mounts.

    Read the article

  • How to install windows 7 from scratch on a disk which already contains partitions

    - by rangalo
    Hi, I have following partitions on a 1 TB disk. 14 GB UNKNOWN recovery partition 100MB NTFS System Reserved partition for Windows 7 448GB NTFS Windows 7 system partition 468GB NTFS Data partition for windows 7 Now because of the problems mentioned in my other question here I got a brand new windows 7 cd and want to install it from scratch after deleting all the extra partitions. But windows 7 installation doesn't give me such options. It refuses to touch the 14GB Recovery and 100 MB (reserved by previous windows 7) partition. Any ideas ? Note: Because of it is a dynamic disks most of the freely available tools refuse to delete the partitions on the disk. regards.

    Read the article

  • RAID P410i and P812 performance issues

    - by Alexey
    I'm having much trouble with I/O performance of HP DL360 server with two RAID controllers - P410i and P812, Windows Server 2008, 36 GiB RAM and 16 x Intel Xeon x5550. The server runs a bunch of tasks producing heavy sequential I/O, and after about 20-30 minutes of intensive work it looks like the tasks are stuck, not using CPU and with enough free memory (so this cannot be a bottleneck). The same tasks were running quite well on the older server (Windows Server 2003, 4 x Intel Xeon, 12 GiB RAM). RAID cache is present, write-cache battery is installed. Cache is configured as 25% readahead/75% writeback. The swap file resides on the logical disk served by P410i and other logical disks are on P812. Can someone tell me what can be the matter of this? Is this a hardware problem or misconfiguration?

    Read the article

  • Hyperthreading vs. SQL Server & PostgreSQL

    - by IanC
    I have read that hyperthreading is a "performance killer" when it comes to DBs. However, what I read didn't state which CPUs. Further, it mostly indicated that I/O was "cut to < 10% performance". That logically doesn't make sense since I/O is primarily a function of controllers and disks, not CPUs. But then no one ever said bugs made sense. What I read also stated that SQL Server could put two parallel query ops onto 1 logical core (2 threads), thereby degrading performance. I have a hard time believing SQL Server's architects would have made such an obvious miscalculation. Does anyone have and data on how hyperthreading on current generation CPUs affects either of the RDBMSs I mentioned?

    Read the article

  • New Seagate GoFlex external hard drive reporting temperature 50C and higher--defective or normal?

    - by rob
    I have a brand-new Seagate GoFlex hard drive, and my SMART diagnostics program (CrystalDiskInfo) is warnings me that the drive is running too hot. The lowest reported temperature I've seen it is 48C, and the highest so far is 56C (all of last week, it was at 50C). According to several sources I've found online, including a Google study, hard disks operating higher than 40C have shorter lifespans. The temperature in the room is usually about 23C (74F), and orienting the drive vertically vs. horizontally doesn't seem to affect the operating temperature. Does anyone else have a GoFlex external desktop drive that runs at 40C or cooler? Is my drive just defective, or is this high temperature common for these drives?

    Read the article

  • Windows Server 2003 Synchronize Not Sticking

    - by lkessler
    We have a Windows Server 2003. It had Raid running on 2 disks. One disk failed and the Raid Controller failed. We replaced the disk and controller and restored everything. No data was lost. The users of that server found that there were a number of directories that appeared empty. We found that from their machine, we could right-click on the directory and select "Synchronize" and the files in the directory would now be visible to them. However, when opening Internet Explorer and browsing the web and ftp'ing to a web site, the files in the directory would vanish. We would have to "Synchronize" them again to get them to reappear. What is going on to cause this need to Synchronize and then re-Synchronize again? What do we need to do to fix this so that the directories are permanently visible?

    Read the article

  • How to recover from a drive failure in a RAID 5 configuration?

    - by Philip Fourie
    This morning a drive failed on our database server. The drive array (3 disks) is setup in a RAID 5 configuration. While we wait for a drive replacement we are preparing for a recovery strategy. Users are continuing to work on the system, albeit very slowly (don't know why??). How does one install the new drive - will the data for this drive automatically be rebuilt from the parity or is there another process we should follow? Edit: This is a hardware RAID controller. (Thanks for the answers so far, appreciated)

    Read the article

  • Alternative to Windows Home Server (WHS) backups

    - by Adam Tegen
    Since Microsoft announced the end of life for WHS, are there any alternatives? Specifically, I am interested in recovering from a catastrophic disk failure with WHS. For example, this is my ideal scenario when a desktop hard-drive fails (has a bad virus, etc): Install a disk of the same size or greater Boot the desktop with the Recovery Disc Point the recovery application at the WHS Pick the machine, the drive(s) and the date of the backup Have a couple beers Reboot to a working machine as if nothing happened. I would need to slap multiple disks in the machine without raid. It sounds like LVM will work here. It would be nice, but not required to have de-duplication of files when multiple machines are backed up. (Single Instance Storage)

    Read the article

  • RAID10 without write-back cache = horrible write performance?

    - by Harry Mexican
    I have just provisioned a dedicated server on singlehop. I'm running it through some tests to know what to expect performance-wise. On the I/O side (with 4 1TB disks in RAID 10) I get: write-cache disabled 200 MB/s read throughput 30 MB/s write throughput I thought that was really low compared to my desktop HD which gets 150-150 or so. So I had a chat with them and they suggested enabling the write cache. New results: write-cache enabled 280 MB/s read 260 MB/s write which is great and all but means I'd have to add a BBU for an additional monthly cost. Is it normal for the write throughput to be 1/4 of a regular drive on RAID10, if you don't have write cache? It almost feels like its intentionally bad to force you to pony up for the BBU. I'd be happy with normal non-raid performance of 150/150.

    Read the article

  • How to persist changes to instances in the cloud.

    - by Peter NUnn
    Hi folks, I must be missing something here, but can someone clue me in on how to persist changes (such as software installs etc) on machines in the cloud (either EC2 or my own Eucalyptus cloud). I have instances running.. can attach extra disks to them etc., but every time I terminate the instance, all of my changes are lost the next time I run them. Now, this sort of makes sense in that the instances are virtual, but, there must be some way to make these changes persist. I'm just missing how its done. Thanks. Peter.

    Read the article

  • Mac Disk-Utility input/output error

    - by Michelle
    a couple of other people have posted about this but my specific problem has not been addressed. For months I have been backing up DVDs and home movies with no problem then all of a sudden I get an "input/output" error. Yes I have cleaned the disks. Actually I have tried 8 different ones - they are not all bad so its obviously my computer. I have done a scan and cleaned up the HD a bit just in case but nothing is helping. I don't want to download other programs since this one works but seems to be having a problem. Any ideas?

    Read the article

  • How can I connect a SAS drive to USB? [closed]

    - by dave
    I have a Dell T710 with Seagate Cheetah 15k.7 SAS disks. If the T710 motherboard dies, I'll need to resort to one of my nightly off-site backups and salvage the journal/logfile from the SAS disk to bring the backup bang up-to-date. I need a way of reading the healthy-but-inaccessible SAS disc that does not depend on the only SAS-capable machine I have to hand. So I bought: SAS to SATA Adapter and: USB 2.0 to SATA Adapter with Power ...so that I could read the SAS drive via USB. I can plug it all together just fine. The chain looks like: USB - SATA - SAS. But the drive does not spin up and the computer doesn't even acknowledge anything being attached by USB. Is there a cheap external enclosure I can buy for SAS drives? I can't believe these USB to SATA adapters are everywhere but the USB to SAS adapters are almost non-existent...

    Read the article

  • What causes "All-in-one USB Card Reader" to create 6 drives that always appear in Disk Management?

    - by tim11g
    I installed a "All-in-one USB Card Reader" to read SD cards and other media. It has caused six new drives to appear in Disk Management with six new drive letter assignments. These drives and letters are always present, even when there are no cards in the reader. When unused, they are labeled "No Media". Why does this multifunction reader cause these phantom Disks to appear and consume drive letters? Every USB port can (and does) allow removable media to be mounted and assigned a drive letter, and the drive letter assignment "disappears" when the USB drive is removed. Why are these card reader's drives and letters staying allocated permanently? Is there anything that can be done to make the slots work like a typical USB drive? (The reader is in fact connected to USB).

    Read the article

  • Windows 8.1 upgrade created a second recovery partition. Can I remove the original?

    - by Dave S
    Windows 8.1 upgrade created a second recovery partition. Can I remove the original? Prior to this the partitions were Recovery, EFI, OS(C:), Data(D:). After the upgrade partitions are Recovery, EFI, OS(C:), Recovery, Data(D:). The first Recovery partition is 1023MB the second is 350MB The "Create a system image" tool selects the EFI, OS(C:), and the second 350MB Recovery partitions. The first 1023MB Recovery partition is not listed, I have to "assume" it is now redundant. The factory (HP) Recovery Partition was removed using the HP provided tool after creating recovery disks, and the D: partition created months ago.

    Read the article

  • There are no drives listed during windows 7 system recovery

    - by Kragen
    I'm trying to use the Windows 7 system recovery disk to repair a boot sector, however I'm finding that when I boot the recovery disk, it doesn't list / mount any of my disk partitions, and so I can't perform the recovery. The partitions are all NTFS formatted, and the drivers used to read the disks all seem to be fairly straightforward Microsoft drivers, so I shouldn't need to load any extra drivers to see my partitions (its a Dell Latitude D530) Diskpart correctly lists the partitions (complete with labels) - it just that when I attempt to switch to that partition it gives me the "This partition does not contain a recognised file system" message. Has anyone got any idea how I can work out why my partitions are not visible?

    Read the article

  • Very slow disk performance on Dell PowerEdge 2950 w/ PERC 6/i running RAID 10

    - by vocoder
    I recently set up a server running Ubuntu 10.04 LTS on a Dell PowerEdge 2950 server - it has 6 500 gb 7200RPM SATA drives setup in a RAID 10 config. I am seeing extremely poor disk performance - the RAID array reports all disks are normal and using MegaCLI, it looks like the BBU is fine. hdparm -tT /dev/sda reports: Timing cached reads: 90 MB in 2.05 seconds = 43.96 MB/sec Timing buffered disk reads: 24 MB in 3.11 seconds = 7.72 MB/sec So as you can see, it takes forever to something as simple as an apt-get upgrade and even logging into the server. How do I go about troubleshooting what is causing this? I upgraded the firmware on the PERC 6i RAID controller to the latest, but didn't see any improvements.

    Read the article

  • differencing disk opinions

    - by troth
    I've read about the performance issues with dfferencing disks but I still think there is a solid place for them and thats the os boot partition. If I'm going to have 20 vms on a csv based volume I don't won't to waste the 20+ gigs per guest just for the os boot. If I get a good base disk with all of the most used applications installed and have the pagefile located somewhere else I don't think the delta's would be that great thus it should not create a performance issue. Also in a SAN based csv volumes does it make any sense in having the pagefile go to a seperate csv volume? Any opinions on this? thanks

    Read the article

  • Time Machine for Windows

    - by Kevin L.
    A simple Google search for "Time Machine for Windows" results in a flurry of different little apps. But instead of relying on forum anecdotes and advertisements, I call on the much wiser Super User beta community for some depth on this one. Having Time Machine running on Leopard is like a warm, fuzzy blanket of comfort that I never got with RAID, rsync, or SyncToy on Windows. I'm not asking the community what the "best" backup software for Windows is, but instead: Is there any true Time Machine clone for Windows, one that includes as many of the following as possible: Completely transparent, "set-it-and-forget-it" backup Incremental backups (changes only) for every hour for a day, every day for a month, and every week until the backup disk is full Ability to rebuild from this backup disk in case of main drive meltdown (the backup doesn't have to be bootable; neither are Time Machine disks) Extremely easy to use UI (target user == wife). Bonus points for a beautiful UI

    Read the article

  • RAID-0 problem with a Sony sporting a new HDD

    - by redrock
    Sony Windows 7 PC. Originally had 2 x 300 Gb HDD. One HDD completely pancaked so have replaced with a new 500 Gb HDD. When both drives are connected the 300GB doesn't appear to be recognized as a 300Gb HDD as a separate entity. BIOS sees it but the operating system only sees a total of 465GB of HD space. When both disks are attached under disk management it shows one 465Gb as RAID 0 and the new drive as STxxxxxx 465Gb. My question I guess is what should I see in total HDD space and is this configured correctly as I thought I would see 2 separate drives 1x500Gb and 1x300Gb. My customer insisted that prior to the HDD crash he saw 2 drives both registering as 300Gb (a c: and d: drive).

    Read the article

  • SBS 2003 R2 install errors

    - by piagetblix
    I just installed sbs2003 R2 as a learning lab box on my home network. The install seemed to take quit a while with a bit of shuffling between disks 4 and 5. The continue setup checklist showed the red x besides the exchange install and finally gave me an error about the "schema needing extending run, forest /adprep" and told me at the end that the install did not succeed. After a longer than normal reboot it comes up and seems to be OK. What logs do i need to check to see what went wrong? Should i reinstall and try burning the iso's to new media? cheers

    Read the article

  • Home server - HP Proliant Microserver - Software and setup - OS on USB stick?

    - by Lloyd Watkin
    I've just purchased a HP ProLiant Microserver for home use. I want to set up with web server, samba shares, the usual stuff. My question is really about system setup. It has an internal USB socket so I've attempted to install a copy of Fedora 14 onto it. I turned off X/Gnome, but it still ran like a pig. I've now put the OS on one of the internal disks (250Gb, 7200rpm), but I was wondering if there was a way to utilise the internal USB to give me better power-saving allowing the hard drives to be shut down when not in use. How would you set this server up? I'd rather not go to the extra cost of an SSD right now, but if that's the best way then so be it.

    Read the article

  • FreeBSD: Samba performance over GBit-Ethernet

    - by Axel Gneiting
    I'm using a FreeBSD NAS with RAID-Z. I can read ~300MB/s from the ZFS disks to /dev/null on the box, but only get about 50MB/s over GBit-Ethernet with SMB to Windows 7 (Samba 3.5.6). Both systems have Intel-PCIe-NICs and are connected directly. Samba is configured to use AIO and I already tried to tune TCP/IP: kern.ipc.maxsockbuf=16777216 net.inet.tcp.sendspace=1048576 net.inet.tcp.recvspace=1048576 net.inet.tcp.sendbuf_max=8388608 net.inet.tcp.recvbuf_max=8388608 net.inet.tcp.delayed_ack=0 Any ideas what's causing the bottleneck? I think the link should handle 100 MB/s easily.

    Read the article

< Previous Page | 39 40 41 42 43 44 45 46 47 48 49 50  | Next Page >