Search Results

Search found 9235 results on 370 pages for 'disk cloning'.

Page 261/370 | < Previous Page | 257 258 259 260 261 262 263 264 265 266 267 268  | Next Page >

  • Brand new Seagate HDD has high raw read error rate

    - by kpax
    I've just purchased a brand new Seagate ST31000524AS 1TB HDD. Manufacture date shows as January 2012 (yes that's as new as new can get), so must be one of the new batches from the post-flood Thailand. Anyway, I downloaded a copy of Active Hard Disk Monitor tool to check the S.M.A.R.T. parameters and I find the parameter Raw Read Error Rate is very low. Should I be worried? Will this rectify over time? This hdd is just 7 hours old; what gives? Edit: I meant high raw read error rate - Title updated accordingly

    Read the article

  • Putting our OLTP and OLAP services on the same cluster

    - by Dynamo
    We're currently in a bit of a debate about what to do with our scattered SQL environment. We are setting up a cluster for our data warehouses for sure and are now in the process of deciding if our OLTP databases should go on the same one. The cluster will be active/active with database services running on one node and reporting and analytical services on the other node. From a technical standpoint I don't see an issue here. With the services being run on different nodes they shouldn't compete too heavily for resources. The only physical resource that may be an issue would be the shared disk space. Our environment is also quite small. Our biggest OLAP database at the moment is only about 40GB and our OLTP are all under 10GB. I see a potential political issue here as different groups are involved but I'm just strictly wondering if there would be any major technical issues that could arise from this setup.

    Read the article

  • My 4 GB microSD card only allows me to use 1 GB

    - by James Litewski
    My phone came with a 4 GB microSD card. On the card it lists that 3 GB goes to Muve Music which is Cricket's music program, and I get 1 GB... Well, I don't pay for Muve Music, so why waste the space? I thought I'd be able to simply buy an adapter and reformat the microSD card to get the full 4 GB; but that wasn't the case... I could only find the 1 GB partition on the card. I even tried reformatting the disk, but I had no luck. How can I get the full 4 GB? BTW, I'm running Mac OS X v10.7 (Lion).

    Read the article

  • DDRescue on Windows or another options like DDRescue for Windows

    - by Frank Thornton
    I have a drive with failed sectors ect... I can't image it with Acronis as it hangs. I can't with Knoppiz it hangs. CHKDSK hangs. I want to use DDRescure but I don't have any Linux boxes running at the moment. I could do is in a VM but that seems like it would be slow and problematic? Are there any ways I can data recovery this disk from my Windows machine or is there an ideal way to work with DDRescue on Windows?

    Read the article

  • Ways to remotely reboot a Linux system

    - by dualed
    I had a remote server running Debian Sarge that experienced some HDD failure and I meant to reboot it hoping that fsck could repair the errors automatically. I eventually drove out there and replaced the faulty disks... But I was wondering: What other ways are there to force a Linux system to reboot that do not require hard drive access? What I had tried: shutdown -r now Did not work, as shutdown is a program that would have to be loaded from disk, the error shown in the terminal was bash: /sbin/shutdown: Input/output error init 6 same as above telinit q same as above kill -2 1 This did not print an error, but did not work either. (However, it is possible that the Sarge init did not implement SIGINT, the sarge manpages did not mention it. So it could work in a more recent version of Debian) This guide on PCFreak.net. However, this failed at sysctl, which was not in memory either.

    Read the article

  • Resize ntfs system partitions with GParted?

    - by ane
    Trying to resize 2 ntfs system and boot partitions (windows 2003 server) using GParted. Goal: Resize D: (/dev/sda1) to ~850G - this is the boot drive with D:\ntldr, boot.ini, etc. Resize C: (/dev/sda5) to 100G - this is the system drive with C:\windows Tried resizing /dev/sda5 first and got the chkdsk error shown in screenshot #2. (You must run chkdsk /f). Have already run chkdsk /f on C: multiple times with no bad sectors or errors found. Have also run multiple chkdsk /f's on the underlying hard disk multiple times and rebooted way more than a couple times with the same error. How do you force gparted to ignore this error and resize? I found there is --force option to ntfsresize but don't know how to get the GParted ISO live CD to use it. How do you move the unallocated space so an extra ~750G is to the right of /dev/sda1 (D:), and an extra 10G to the right of /dev/sda5 (C:)

    Read the article

  • Why does diskpart set the volume attributes on all volumes?

    - by Nick
    I was trying to migrate a Win7 OS from a HDD to a SSD. I've created 2 partition with 1024KB offset, with diskpart: 100MB System Reserved and a 60GB for C:. I've cloned their contents using Easeus Disk Copy. I've loaded the Windows 7 Boot DVD, and wanted to use diskpart to drop the letter for the System Reserved partition and make it hidden. select volume 0 detail volume attribute volume set nodefaultdriveletter attribute volume set hidden These 2 attribute set commands actioned on both volumes (0 and 1, MSR and C:) instead of the selected one, and viceversa. I've tried to clear these attributes from volume 1, but it cleared them also from volume 0. Why does DiskPart have this behaviour?

    Read the article

  • Create a Windows 7 System Repair ISO?

    - by MaDMaD Mad
    According to the Microsoft website I can run "Create a system repair disc" from the start menu to burn a recovery Disc. However, there are many situations (including not having a disk drive) where this cannot be done. My goal is to create a bootable USB drive according to this guide but their method to do so without burning a physical disc includes downloading an ISO from "somebody's" website. I don't know who "somebody" is and I feel it is reasonable to not boot something I can't verify as being untampered with. To the best of my knowledge, Microsoft does not host a recovery disc ISO for download as this would be the easiest solution. So my problem is that I'd like to circumvent Microsofts "burn to disc" option with a "burn to ISO" through any reasonable means.

    Read the article

  • Problem opening password encrypted .docx file on Word 2003

    - by molecule
    Hi all, I am having a problem opening a .docx file on my Word 2003. I have installed the Compatibility pack for 2007 but when i try to open this particular file, I receive the error "Word experienced an error trying to open the file. Try these suggestions. 1. Check the file permissions for the document, 2. Make sure there is sufficient free memory and disk space, 3. Open the file with the Text Recovery converter. I do not think it is any of the errors as I am able to open it on a different PC with Word 2003 as well. I also do not have any issues opening any non-password encrypted .docx files. Has anyone experienced the same issue? Most posts on the internet relate to "open and repair" but as mentioned, I am able to open this file on another PC without any problems. Any advice is greatly appreciated. Thanks, George

    Read the article

  • How to get external 2.5 inch HD enclosure get recognised

    - by fireBand
    I am completely out of ideas. I have 2.5inch HDD from my old laptop, I have tried with 2 different HDD enclosures but my desktop does not recognize it. Not even showing in the disk management or BIOS. The HDD works fine if I use in a 3.5 inch enclosure with external power. I have tried 3 different Y USB cables. I have also got a 7 port self powered 2A USB hub. I have updated my chipset drivers as well. The problem is with my desktop , the enclosures work fine with my new laptop. Anything else that I should try? Thanks in advance.

    Read the article

  • Can't mount home after trying to resize (bad geometry: block count exceeds size of device).

    - by Lynn
    This is on a fresh computer (super computer actually). It got to me with 15T on the home mount and 50G on the root. I tried allocating 7T to root and resizing (since I'm putting a local yum repo on this machine as it has no internet access nor will it ever). I tried following the instructions here: Centos 6.3 disk space allocation but something went wrong and the home won't mount again. Instead I get from dmesg | tail: EXT4-fs (dm-2): bad geometry: block count 4294967295 exceeds size of device (1342177280 blocks) df -h nets this output: Filesystem Size Used Avail Use% Mounted on /dev/mapper/VolGroup-lv_root 7.0T 3.6G 6.6T 1% / tmpfs 190G 216K 190G 1% /dev/shm /dev/sda1 485M 38M 422M 9% /boot I didn't have any files on /dev/mapper/VolGroup-lv_home. Will simply running mke2fs fix it to be mountable? What sort of options should I run it with. I've never resized volumes before or used mke2fs. I don't want to make this mess worse.

    Read the article

  • Intermittent USB 3.0 access - How do I troubleshoot?

    - by Billy ONeal
    I've got a WD Passport hard disk with "superspeed" USB 3.0 support. When I use my USB 3.0 flash drive (this is a Lenovo X220 laptop), USB 3.0 consistently works. But when I use the passport drive, almost without fail the connection drops to USB 2.0. Touching the cable seems to immediately trigger the problem, but it seems to happen on its own though. I've got another cable on order right now... but is it likely the cable's the issue here? Is there anything else I can check?

    Read the article

  • Is this a valid backup strategy for MongoDB?

    - by James Simpson
    I've got a single dedicated server with a MongoDB database of around 10GB. I need to do daily backups, but I can't have downtime with the database. Is it possible to use a replica set on a single disk (with 2 instances of mongod running on different ports), and simply take the secondary one offline and backup the data files to an offsite storage such as S3 (journaling is turned on)? Or would using master/slave be better than a replica set? Is this viable, and if so, what potential problems could I have? If not, how do I conceptualize this to work?

    Read the article

  • Pervasive database backup

    - by Steven
    I'm looking for the best way to backup my pervasive database. I've read the documentation but still have a few questions. It appears that Continuous Operations method only allows me to backup the entire database? So I'd do butil -startbu @filelist, then backup the entire database (copy, rsync, etc), then run butil -endbu @filelist. Looking through the documentation I don't see a way to get transaction logs out of this method; like I would do for MSSQL (BACKUP LOG ACCT TO DISK) or Postgres (archive_command). With rsync, it might be feasible to still do this every 15 minutes. The Archival Logging method means I would have to occasionally stop the database to get a full backup, which is acceptable for me. But can I copy the log files off of the server every 15 minutes, ie log shipping? Thank you.

    Read the article

  • Computer Locks Up Completely When Resuming From Standby

    - by kcoppock
    Okay, so I'm having a strange issue lately with my desktop, and off the top of my head I'm not sure what the problem is, so I'm hoping for some ideas on how to track down the cause. First, my PC: AMD Phenom 9600 Quad-Core 2.4 GHz 4 GB DDR2 160 GB boot HDD 1 TB & 120 GB secondary HDD Nvidia GeForce 8600 GTS Windows 7 Professional 64-bit I have my power settings as follows: Turn off the display: 10 minutes Turn off hard disk: 20 minutes Sleep after: Never After a while, I'll come back to use my computer again, wiggle the mouse, and the screen comes back on like normal, but the desktop is unresponsive. I can move the cursor, but nothing responds to clicks or keyboard entry. If I click a few times with the mouse, the computer eventually locks up completely and the mouse cursor stops moving as well. After I restart, everything works just fine. Is there a log you recommend checking or really any suggestions would be welcome at this point. Thanks!

    Read the article

  • Anybody know how to get podcasts sorted correctly on a Zune?

    - by Will
    If I could understand the logic behind it I could actually manage to get my stuff sorted correctly on the device. It doesn't matter how things are sorted on disk, or in the software itself. As items are added to the sync list they end up sorted in some weird manner which, quite frankly, brings out an anger in me that scares me. A primitive, misshapen and unthinking anger. Kind of like the lumpy guy in Altered States. Frankly, I'm afraid one day its going to burst forth and I will track down the entire Zune team and make them pay for their failures. Help me prevent this from happening. Help me understand what I need to do to ensure my podcasts end up in the order I wish them to be in, not some obscure random order that makes no sense.

    Read the article

  • How long do uploaded files stay in the tmp folder in Linux Ubuntu?

    - by Jean-Nicolas Boulay Desjardins
    I am building a web application where my users will be able to upload files. After the files are uploaded I need to send the files to two other servers, and after they will be deleted from the server where they were just uploaded to. I am wandering is it a good I idea to keep the uploaded files in the tmp/ folder the time the uploaded files are sent to the other two servers or should I move them to another folder incase they get deleted? I am also wandering because I would like to know if I have to build a cron script to get rid of the files that have been transfered to the other servers so that I get my disk space back.

    Read the article

  • Windows 7 Public Documents Folder - how to move it to drive d:

    - by Bazza Formez
    Hi, I've just bought a new pc, and it is running Windows 7. I'm wanting to set up all of the documents folders such that they point to folders on a second drive partition (ie. disk D:) so that I have a nice separation of os from docs. I have managed to do this for each user on the machine (by opening up properties for documents folder and changing location). Hwoever, this procedure doesn't seem to work for the public documents folder. I cannot change the location for that one (even though I have administrator privileges). It is greyed out and cannot be changed. Any ideas ? Thanks!!

    Read the article

  • Shared volume for data (multiple MDF) and another shared volume for logs (multiple LDF)

    - by hagensoft
    I have 3 instances of SQL Server 2008, each on different machines with multiple databases on each instance. I have 2 separate LUNS on my SAN for MDF and LDF files. The NDX and TempDB files run on the local drive on each machine. Is it O.K. for the 3 instances to share a same volume for the data files and another volume for the log files? I don't have thin provisioning on the SAN so I would like to not constaint disk space creating multiple volumes because I was adviced that I should create a volume (drive letter) for each instance, if not for each database. I am aware that I should split my logs and data files at least. No instance would share the actual database files, just the space on drive. Any help is appretiated.

    Read the article

  • Raid on ICH9R chip set

    - by user500982
    Hi Im looking at buyign this MB: http://www.supermicro.com/products/motherboard/ATOM/ICH9/X7SPE-HF-D525.cfm Im wondering though if the chipset will support the raid configuration I need. Im looking to configure the following arrays: raid array 1: 2 2TB Disks in Raid 0 raid array 2: 2 2TB Disks in Raid 0 raid array 3 (not actualy an array): 1 300GB Disk not in raid, to be used for OS and boot. So in total there would be 5 drives and the board supports 6. so im good when it comes to connections. However I have herd some chip sets only support one raid array (volume). so either all drives are individual, or are in the array. I must have 2 sperate raid arrays independent of each other, and a 5th drive not in any array. Anybody know if this setup will work? Thanks, -Stewart

    Read the article

  • Cloud based backup solutions based on open standards?

    - by Rick
    I am looking for a solution to backup and consolidate important media from a couple Windows laptops and Mac laptop. I would like a solutions that based on open standards, so my data isn't trapped by proprietary formats and proprietary protocols. I would like the ability to switch clients or change providers in the future. For example, something like Jungle Disk plus S3 sounds like a great option. However, I am having trouble confirming how or if this can be setup meeting this criteria. Are there any real or de-facto standards for treating S3 as a filesystem? If so, what Windows and Mac clients support these standards?

    Read the article

  • Feeding the kernels entropy source from other machines and/or increasing its maximum size

    - by David Spillett
    We have has a little trouble with a small box that acts as a VPN end-point and mail relay for our network, caused by the available entropy for /dev/random being too low (which causes TLS connection attempts by exim to fail). The machine doesn't do anything else, so the normal feed into the entropy pool (interrupt timings from things like disk access) is not enough. As a quick hack I've set a looping script that reads from /dev/hda at a couple of Mbyte/sec which keeps it topped up. Other than buying a hardware RNG, is there a clean way of piping data for entry from elsewhere, such as a copy of the data our file server uses for its entropy source? I've spotted several tips for using rng-tools to feed it from /dev/urandom on the same machine but that "feels dirty". Also, is it possible to increase the maximum pool size? It currently seems to max out at 3585.

    Read the article

  • How can I diagnose cache misses when using Apache as a reverse proxy?

    - by johnstok
    I have set up Apache 2.2 as a reverse proxy with the following configuration: # jBoss proxying ProxyRequests Off <Proxy *> Order deny,allow Allow from all </Proxy> ProxyPass /foo http://localhost:9080/foo ProxyPassReverse /foo http://localhost:9080/foo ProxyPassReverseCookiePath /foo /foo # Reverse proxy caching CacheEnable disk /foo # Compression SetOutputFilter DEFLATE BrowserMatch ^Mozilla/4 gzip-only-text/html BrowserMatch ^Mozilla/4\.0[678] no-gzip BrowserMatch \bMSIE\s(7|8) !no-gzip !gzip-only-text/html DeflateCompressionLevel 9 Header append Vary User-Agent env=!dont-vary However, in a number of cases where I expect a cached response to be returned the request is sent through to the origin server at localhost:9080. Responses have a HTTP Vary header of 'Accept-Encoding,User-Agent' which is to be expected given the mod_deflate configuration. How can I determine why Apache is unable to serve a response from the cache?

    Read the article

  • My WD passport detected but not shown on my computer

    - by Rakesh Patel
    1TB WD harddisk is detected in laptop but drives not visible on my computer. Under disk managment drives shown without letters. Right clicking on these drives give 2 options: Delete volume Help. The same harddisk was running well on the same laptop a day before. It is also showing full size of hard drives, Healthy (Primary Partition) etc. (this happened after harddisk usage by guest). Harddisk is new only (about 5 month usage) Please suggest: Is there hardware related problem inside harddisk, if so is it repairable? Is there any option to recover data?

    Read the article

  • Solution to Manage and Monitor (Ubuntu) Machines

    - by Elmar Weber
    I'm looking for a tool like Canonical (system management and monitoring for Ubuntu) that is Open Source and free. The goal is to manage a dozen or so KVM machines for private testing purposes. I know of puppet and munin or RHQ as separate tools to manage and monitor, but I'd prefer something integrated. Any tips? Basic requirements would be: system package management and update (individual selection for each managed node) configuration of basic system services (Users, NFS, cron, ideally also Apache) monitoring (charting of system resources, disk, io, memory, etc) and alerting, ideally a default configuration with sensible values for alerts

    Read the article

< Previous Page | 257 258 259 260 261 262 263 264 265 266 267 268  | Next Page >