Search Results

Search found 1864 results on 75 pages for 'raid 1'.

Page 58/75 | < Previous Page | 54 55 56 57 58 59 60 61 62 63 64 65  | Next Page >

  • How do I built a DIY NAS?

    - by Kaushik Gopal
    I'm looking for good, detailed instructions on how to build a DIY NAS (Network Access Storage). I'm planning on doing it cheap (old PC config + open source software). I would like to know: What hardware I need to built one What kind of hard-drive setup I should take (like RAID) Or any other relevant hardware related advices (power supply, motherboard etc...) What software I should run on it, both what OS and software to manage the contents effectively So the NAS is recognizable and accessible to my network I can make sure my Windows computers will recognize it (when using Linux distro's) I can access my files from outside my network I already did a fair bit of searching and found these links, but while these links are great they delve more on the hardware side. I'm looking for more instructions in the software side. Ubuntu Setting up a Home NAS DIY NAS Smackdown How to Configure an $80 File Server in 45 Minutes FreeNAS Build a NAS Device With an Old PC and Free Software Build Your Own NAS Device

    Read the article

  • Win 2008 R2 Server Not Recognizing Second Hard Drive

    - by Brian
    Hello, I just purchased a Dell server, which has two hard drives and no RAID setup. I can only currently see one hard drive... not sure how to get it to recognize the other, as I thought being a new machine that wouldn't be an issue. It has Windows Server 2008 R2 that I loaded on. I'm a n00b to all of this so I'm not sure why this is failing to work... Any help appreciated. Thanks.

    Read the article

  • for what reason live USB linux images are not 100% persistent?

    - by gcb
    I understand that during CD-R times non-persistence had a purpose, but what's the purpose now that pretty much everyone uses USB flash drivers? not to mention USB3 sticks are pretty much 4x faster than my HD raid. I'm writing this while taking a break going over the linux from scratch guide... And i'm still baffled that this is not the norm already with all live images. So, Is there any reason (besides historical) that i'm missing and that will bite me after I finish this ext3-rw image?

    Read the article

  • VSS error 12293 after system disk clone (Win2003)

    - by carlpett
    Hi! After cloning a windows 2003 installation from a single drive onto two mirrored drives using Acronis Disk Director, VSS no longer works, filing events 12293 and 7001 when trying to use backup tools, and additionally giving error 0x8004230f when accessing the Shadow copy tab of disk properties. I've google-researched this quite throughly, and found a suggested fix[1]: replacing the MBR signature of the disk. This would cause windows to invalidate old shadow copy information, which supposedly would make it all work again. However, I am a bit nervous over this... Is there a possiblity of messing this up somehow, because of the mbr originating from a single disk install, and now residing on a raid mirror? Has anyone here had this problem and solved it? This method or another? [1] http://kb.backupassist.com/articles.php?aid=2971 (under header Resolution 2)

    Read the article

  • rsync error: some files/attrs were not transferred

    - by Daniel Ball
    Using rsync(ubuntu) and a DeltaCopy server on W2K3 to back up some of the data on the file server before I migrate from W2K3 to Ubuntu server. After it completed I ran a dry run just in case something had been missed or changed ... I got the following: sudo rsync -az -n 198.3.9.25::Music /mnt/raid/music [sudo] password for daniel: file has vanished: "?????\#267????" (in Music) file has vanished: "????????" (in Music) ... rsync error: some files/attrs were not transferred (see previous errors) (code 23) at main.c(1526) [generator=3.0.7] I just want to make sure I'm reading it right, that somehow there are files on the receiving end that aren't on the sending?

    Read the article

  • Format (remove) HDD volume that is visible in Windows 7 Disk Management but not diskpart.exe

    - by EntropyWins
    I'm trying to get iRST working on a SSD I installed in my lenovo u410. As part of that process, I created a hibernation partition and was fiddling around with RAID/AHCI settings. I managed to make my computer unbootable. No sweat, I just restored it with Lenovo's 1 key system. Now, however, I can't do anything with that hybernation partition! I can see it: (It's the 7.81 GB partition). But when I try to delete it in Diskpart.exe to reclaim the space and try the formatting again I only see this: I can't do anything with the partition in Disk Management either. Right clicking only shows the 'help' option. Can anyone suggest a way to edit these partitions with windows or, at least, reccomend a program that might help me fix this? Note, I'd rather not delete the 16 GB OEM partition that I believe holds the backup for this computer.

    Read the article

  • Dedicated GPU in Dell PowerEdge C1100

    - by Eli Gundry
    We recently purchased a Dell PowerEdge C1100 off lease with the initention of using it for graphics processing. We installed an AMD HD 7000 series GPU in it that runs off of board power and it sends video to the display. That said, the video is very choppy, leading us to belive that the onboard video is doing all the processing and sending it to the card. Is there any way to either disable the VGA on this server or tell the OS to only use the dedicated card. More info: The server is running REHL 6.4 The graphics running the proprietary AMD drivers The video card only works in OS and does not show the BIOS on boot (we know that it's impossible to change this) Any ideas, guys? Update We are now thinking that the GPU is doing the graphics processing, but not working at the full speed of the PCI bus. Which is odd, because it is an x16 slot, but probably optimized to use a RAID card (if that makes any sense). Is there any way to remedy the choppy graphics on this server?

    Read the article

  • Reduce the I/O priority of Windows Backup (Windows Server 2008 R2)

    - by HelloSam
    I have a PostgreSQL running on Windows Server 2008 R2 x64 box. And I have scheduled a backup everyday from the RAID 1 DB disk to a dedicated standalone disk. They are SAS 15k on Dell PERC 6i. I am using the built-in Windows Server Backup for purpose. The problem is, whenever the backup process is kicked in, the database performance is hogged. I would say almost a 10x of performance reduction. From the resource monitor, the disk queue is in the double digit range when backing up, and less than 1 during the day. The disk activity is like ~30-50MB/s during backup, so I guess the hardware is acting normally, though wbengine.exe takes up most of the portions. I think reduce the IO priority of the backup process would be an answer, but I couldn't find a way to. Tuning process CPU priority does not seems to help.

    Read the article

  • Unix Server Partitioning & Filesystem Layout

    - by user1717735
    There's a lot of contradictory information about Unix server partitioning out on the internet, so I need some advice on how to proceed. So far, on the servers I in our test environment I didn't really care about partitioning and I configured a single monolithic / plus a swap partition. This partitioning scheme doesn't seem like a good idea for our production servers. I have found a good starting point here, but it seems very vague on the details. Basically I have a server on which I will be running a basic LAMP stack (Apache, PHP, and MySQL). It will have to handle file uploads (up to 2GB). The system has a 2TB RAID 1 array. I plan to set : / 100GB /var 1000GB (apache files and mysql files will be here), /tmp 800GB (handles the php tmp file) /home 96GB swap 4GB Does this sound sane, or am I over-complicating things?

    Read the article

  • Invalid BOOT.INI (dual boot XP with 7)

    - by Muxa
    I had Windows XP x64 as my main system, and i also had a second partition with Windows XP x64. Both booted from first partition (C:) I then installed Windows 7 Ultimate on the first partition. I've added NTLDR using BCDRDIT. I've also copied NTLDR, NTDETECT.COM and BOOT.INI onto the drive where XP remained. However then i try to boot into Windows XP x64 i get Invalid BOOT.INI file Booting from c:\windows\ NTDETECT failed I found instructions on how to fix it using a boot disk, however the partitions are on a software RAID. I've tried to boot from a customized XP CD with the drivers, however it does not offer me a Repair option for some reason - just setup. Partitions that i have:'= System Reserved Main (Windows 7) Secondary (Windows XP x64) Here's the contents of my BOOT.INI: [boot loader] timeout=30 default=multi(0)disk(0)rdisk(0)partition(3)\WINDOWS [operating systems] multi(0)disk(0)rdisk(0)partition(3)\WINDOWS="Windows XP Professional x64 Edition" /fastdetect

    Read the article

  • Win 2008 single server development environment (architecture)

    - by Tommy Jakobsen
    I have a few questions about a test development environment that I’m setting up on this server: Intel Core i7-920 Quadcode incl. Hyper Threading 8 GB DDR3 RAM 2x 750 GB SATA-II (probably software RAID 1) The server is going to support max 5 users, maybe 10 when stressed. I was hoping that I could run all the following products on the same server: Windows Server 2008 R2 x64 w/ IIS SQL Server 2008 x64 (R2 when released) Team Foundation Server 2010 Sharepoint Foundation 2010 I know this sounds overkill, but remember that this is for development purpose and testing. This is not a production environment. My question if this will be possible at all? Should I run it all on one Windows 2008 installation, or should I run it in multiple virtual environments using Hyper-V? What do you think?

    Read the article

  • optimizing file share performance on Win2k8?

    - by Kirk Marple
    We have a case where we're accessing a RAID array (drive E:) on a Windows Server 2008 SP2 x86 box. (Recently installed, nothing other than SQL Server 2005 on the server.) In one scenario, when directly accessing it (E:\folder\file.xxx) we get 45MBps throughput to a video file. If we access the same file on the same array, but through UNC path (\server\folder\file.xxx) we get about 23MBps throughput with the exact same test. Obviously the second test is going through more layers of the stack, but that's a major performance hit. What tuning should we be looking at for making the UNC path be closer in performance to the direct access case? Thanks, Kirk (corrected: it is CIFS not SMB, but generalized title to 'file share'.) (additional info: this happens during the read from a single file, not an issue across multiple connections. the file is on the local machine, but exposed via file share. so client and file server are both same Windows 2008 server.)

    Read the article

  • Best Solution for Load Balancing geographically distributed NFS File Access?

    - by DairyKnight
    I'm trying to find an optimum solution for accessing the NFS file share in my company. We have a central file server in North America and has 30GB~50GB of updated data everyday. And it's very slow for our Europe and Asia branches to access directly. Therefore, I'm trying to setup two replicate servers in those continents. I'm currently using rsync, but wonder if there exists a better solution acts more like a distributed RAID, which allows the user to transparently access the file whether synced or not. And user request will be dispatched to remote server if the file is not yet synced. I'm now looking into DRBD, but it seems not to have the functionality of auto-dispatching requests. Does anyone know if there's a better solution?

    Read the article

  • Natively boot Virtualbox Image

    - by isync
    I am faced with a Windows hardware/software problem left over from another person. It's on me to resolve. It's a mission critical setup. The situation is: I've got a physical server machine with: -Disk C:\ (one disk) containing a basic install of Windows Server 2008 R2, formerly Win Vista Pro, now gone. -Disk D:\ (software Raid) containing a VirtualBox disk image of a configured Windows Server 2008 R2 running SQL Server R2 among others. What shall I do now? Migrate all the stuff from the configured VM to the basic but natively installed C:\ Windows Server 2008 R2 (with the possibility of breaking stuff)? Or, Setting up the machine to "natively boot" the VM with the help of bcdedit.exe (something I've read about, what I've never done, what I don't know of if it works, if it hits performance, or if it is stable for production) For me, being old skool, I am in the process of de-virtualising everything (option 1). But I'd be happy if someone suggests I am ok to go down the "natively boot" route.

    Read the article

  • postgres memory allocation tuning 2

    - by pstanton
    i've got a Ubuntu Linux system with 12Gb memory most of which (at least 10Gb) can be allocated solely to postgres. the system also has a 6 disk 15k SCSI RAID 10 setup. The process i'm trying to optimise is twofold. firstly a single threaded, single connection will do many inserts into 2-4 tables linked by foreign key. secondly many different complex queries are run against the resulting data, using group by extensively. this part especially needs to be optimised. i have four of these processes running at once in order to make use of the quad core CPU, therefore there will generally be no more than 5 concurrent connections (1 spare for admin tasks). what configuration changes to the default Postgres config would you recommend? I'm looking for the optimum values for things like work_mem, shared_buffers etc. relevant doco thanks!

    Read the article

  • Shrink Partition on Production Server

    - by Campo
    SO our production server was only setup with one large partition. I have setup a standby server and properly partitioned it. Now the boss wants the production environment's partition shrunk. It is an HP DL380 G5 We have 4 hot swap drives in a raid 5. How best should I go about doing this. Seems like a bad idea to me. Should I use windows or HP to do the partitioning? What should I be aware of in a production environment? The idea is to put the site (Inetpub) on a separate partition instead of the C: drive. How much downtime should I expect? Is this a terrible idea? Anything else I have missed?

    Read the article

  • Disk full, du tells different. How to further investigate?

    - by initall
    I have a SCSI disk in a server (hardware Raid 1), 32G, ext3 filesytem. df tells me that the disk is 100% full. If I delete 1G this is correctly shown. However, if I run a du -h -x / then du tells me that only 12G are used (I use -x because of some Samba mounts). So my question is not about subtle differences between the du and df commands but about how I can find out what causes this huge difference? I rebooted the machine for a fsck that went w/out errors. Should I run badblocks? lsof shows me no open deleted files, lost+found is empty and there is no obvious warn/err/fail statement in the messages file. Feel free to ask for further details of the setup.

    Read the article

  • Using 2-port LSI 2308-8e card to control 24 SAS HDDs

    - by GregC
    I would like to rely on a RAID-on-chip solution to control 24 SAS hard drives in a direct-attached environment. How would you approach this to get best bandwidth given that I'd like to spend less than $10,000 on the interconnect. I've read that LSI 2308 chip can easily handle 8-drive SSD RAID6 in hardware. I'd like to harness its power to control 24 SAS hard drives over an expander in an external enclosure. Currently I use an Infortrend S24S-G2240 external enclosure. It provides its own controller and expander. I'd like to use LSI 2308 controller for RAID6 somehow instead of the mystery controller in the enclosure. P.S. I tried to create SAS-expander as a tag, but my rep on this site is low.

    Read the article

  • Terminal Services - MS Access Frequently "Not Responding"

    - by jonfhancock
    Exposition: We use a program built in MS Access that I serve via Terminal Services. I just installed a new TS Server with a Quad Core 2.6GHz Xeon, 8GB RAM, and 4 SATA drives in a RAID 0. In installed Server 2008 R2 (64bit obviously). It's only role is TS. The problem: With just a few sessions (under 10), I start getting frequent Not Responding messages in each session. When it happens, the users aren't doing anything particularly taxing, just form navigation and simple insert queries. I can live with some stalls, but it is visually jarring in WS08 because the screen goes gray, and it presents a dialog offering to wait or close with some other options. Questions: Any suggestions for improving performance and reducing hangs? Is it possible to disable the dialog (always wait) and screen graying?

    Read the article

  • Getting BootMgr not found errors repeatedly on Win7 x64

    - by abszero
    So here is the basic configuration of the box: Primary RAID 1 (Mirror, Bootable): 2x 300GB WD SATA drives AMD Phenom Quad Core x64 @2.2 ASUS M3N78 Pro Board 4GB RAM Win 7 Ultimate Additionally, this box is a Host OS for several CentOS Boxes via VirtualBox. The box runs like a champ but, for whatever reason, everytime I restart the machine I get a BootMgr not found error when the box tries to boot. I pop in my Win DVD, select 'Repair Windows' then 'Fix Start Up Problems' and everything works fine...once. When I restart the box again I have to go back through this process. Any ideas on what is going on?

    Read the article

  • Verify server performance

    - by George Kesler
    I'm looking for a quick and SIMPLE way to verify that new servers are performing as expected. The most important metric is disk performance, second is network performance. I’m trying to prevent problems caused by misconfiguration of RAID arrays, NIC teaming etc. The solution should work with both physical and virtual servers. I don’t need sophisticated analysis with different workloads, just one set of benchmarks which I would run against a reference server and later compare to new ones. One problem is that most benchmarks are not giving accurate results when running on a VM.

    Read the article

  • How can I tell if I have PCI Express 2.0 or 2.1?

    - by Stefan Lasiewski
    I am looking at a variety of PCI Express cards, such as a SATA RAID Controller and a Video Card. Some of these cards say they only support PCI Express 2.1, not PCI Express 2.0. I know that my motherboard supports PCI Express 2-something, but the manual doesn't say '2.0' or '2.1'. How can I tell if the PCIe slot on my motherboard is PCI Express 2.0 or PCI Express 2.1? Is it possible to determine the PCIe type from the Windows or Linux commandline? I was under the impression that most PCI Express 2.1 devices are backwards compatible with PCI Express 2.0. Is it possible that the vendor is wrong in saying that PCI Express 2.1 is required?

    Read the article

  • Trying to determine the correct number of XFS allocation groups for postgresql server on Linux

    - by HBlend
    I am running a postgres 8.4.5 server on the linux 2.6.33.7 kernel on an 8 disk raid array with an LSI controller. Most of the tables are around 1GB or less. I know that XFS uses allocation groups (AG) to achieve I/O parallelism. My first question is, does this mean that if two tables are in the same AG, all I/O requests are queued to both of them if either is being read from/written to? If so, I assume I would want to spread my tables across as my allocation groups as possible, correct? Wouldn't this ensure that multiple users querying different tables would get the best performance?

    Read the article

  • Why Buy Hardrives with storage server from a vendor?

    - by Mark
    Hi all, Im just browsing around at storage server's like the Dell MD100/ MD3000 and the Sun J4200 and although the storage server seems reasonable (approx $3000-$4000 AUD) the hard-drives that you buy to go along with them seems exorbitantly expensive. And I'm not sure why. Surely at most they are using good quality RAID level 7200rpm SATA hdd, but even then they are still charging almost 4 times the price. What is the advantage to buying these from them. I can see if one fails then the vendor replacing it is convenient. But at that price you could buy double the amount of hdd and just claim on warranty directly with the manufacturer. It would be much cheaper and you wouldn't be relying on someone else to fix your problems. Is this the case of "you don't get fired if you buy IBM?" mentality or is there some reason I'm not grasping here? Cheers Mark

    Read the article

  • Rough estimate for speed advantage of SAN-via-fibre to san-via-iSCSI when using VMware vSphere

    - by Dirk Paessler
    We are in the process of setting up two virtualization servers (DELL R710, Dual Quadcore Xeon CPUs at 2.3 Ghz, 48 GB RAM) for VMware VSphere with storage on a SAN (DELL Powervault MD3000i, 10x 500 GB SAS drives, RAID 5) which will be attached via iSCSI on a Gbit Ethernet Switch (DELL Powerconnect 5424, they call it "iSCSI-optimized"). Can anyone give an estimate how much faster a fiber channel based solution would be (or better "feel")? I don't mean the nominal speed advantage, I mean how much faster will virtual machines effectively work? Are we talking twice the speed, five times, 10 times faster? Does it justify the price? PS: We are not talking about heavily used database servers or exchange servers. Most of the virtualized servers run below 3-5% average CPU load.

    Read the article

< Previous Page | 54 55 56 57 58 59 60 61 62 63 64 65  | Next Page >