Search Results

Search found 20358 results on 815 pages for 'disk management'.

Page 456/815 | < Previous Page | 452 453 454 455 456 457 458 459 460 461 462 463  | Next Page >

  • Is there a way to mirror two severs on Ubuntu?

    - by Kyle
    I was wondering if it's possible to mirror two servers, like you could upload files to one server and they'd push to the other server, etc. I'm more curious to file mirroring, it doesn't have to mirror package management and setup (But that'd be cool too!)

    Read the article

  • Install xenserver on iscsi target

    - by Ghassen Telmoudi
    Is is know that Xen Server is based on CentOS, and it is fact that CentOS can be installed on an ISCSI target flawlessly, but I wanted to install Xen server on an ISCSI target I couldn't find a way. I already tried the latest version of xen server 6.2, and there is no obvious way who to do the installation without a local disk. Does anybody have an idea about the subject, or did someone know how to do it? Please share your experience about this subject.

    Read the article

  • Allowing Apache in Ubuntu to access files in NTFS hard drive

    - by lyrae
    I have LAMP running in Ubuntu. However, my files are located on a separate NTFS hard drive (/media/shared/mysite/). going to http://localhost gives me a 403 how can i, securely, allow apache to read/write the NTFS disk? 'shared' is currently being mounted when system boots. here's the entry in fstab: /dev/sda1 /media/shared ntfs-3g quiet,defaults,locale=en_US.utf8,umask=000 0 0

    Read the article

  • few questions on clone zilla

    - by user23950
    I'm trying to clone my windows xp installation. If I back it up using clone zilla and the my xp machine is infected by virus/spyware would I also be bringing the whole mess when I try to back it up. Do I need the whole partition/ whole disk if I use an external hard drive to backup. Would the data be formatted on the partition that I choose?

    Read the article

  • system center operations manager (SCOM) role combinations

    - by KAPes
    We are evaluating system center operations manager 2007 R2 product, and would like to know what all roles can be combined onto single server. Like root management server and reporting servers can be on one box or not? my environment is like 450 servers mostly Exchange and Domain Controllers plus few OCS servers.

    Read the article

  • How to tune system settings for mongoDB on Linux?

    - by jsh
    Trying to squeeze a lot out of one question here -- please bear with me. Although the MongoDB man pages make several useful recommendations about system settings like ulimit (http://docs.mongodb.org/manual/reference/ulimit/), and other production factors (http://docs.mongodb.org/manual/administration/production-notes/) they seem mysteriously silent on things like virtual memory and swap settings. The closest we get to a hint is that "...the operating system’s virtual memory subsystem manages MongoDB’s memory..." (http://docs.mongodb.org/manual/faq/fundamentals/#does-mongodb-require-a-lot-of-ram). Running the same job - high writes and high reads on about 10,000,000 records in a single collection -- on my 4-processor, 4GB RAM macbook and an 8-core ubuntu box with 64GB RAM I saw dramatically WORSE read performance on the linux box with factory settings, and could hear the disk constantly spinning, indicating high I/O and presumably swapping. Yes, other things were happening on the box, but there was plenty of free RAM, disk space, etc.; furthermore, I did not see evidence that Mongo was expanding to take advantage of all that free RAM as it is touted to do. Linux box default settings were as follows: vm.swappiness =60 vm.dirty_background_ratio = 10 vm.dirty_ratio = 20 vm.dirty_expire_centisecs =3000 vm.dirty_writeback_centisecs=500 I hazarded some guesses looking at docs and blogs for other types of databases (Oracle, MYSQL, etc.), experimented, and adjusted as below. vm.swappiness=10 vm.dirty_background_ratio=5 vm.dirty_ratio=5 vm.dirty_writeback_centisecs=250 vm.dirty_expire_centisecs=500 I saw some immediate apparent improvements in read time. However, when I ran my test jobs again, read performance continued to be painfully sluggish during heavy writes. Then, I REBUILT the collection from an available data source - and suddenly I can read at 1ms or less per record WHILE doing the write job! So the question is really two-fold: 1) What are appropriate VM settings for MongoDB on Linux? 2) (bonus) Does Mongo do some checking or optimization with the OS while data is being built? In other words, if I have built a large data set with suboptimal VM or I/O settings, does Mongo make assumptions during the memory-mapping process that will fail to take advantage of optimizations down the road? Obviously I don't fully grok memory mapping under the hood (I was hoping I wouldn't have to). Any help appreciated...thanks! -j

    Read the article

  • Resources for a new SysAdmin? (Emphasis on Windows SBS, Exchange, networking and general SysAdmin in

    - by 80bower
    I've recently taken over management of a Windows 2003 Small Business Server and network for a small, less than ten person company. I have some (antiquated) sysadmin experience, but I've little experience with Exchange. The documentation of the existing infrastructure leaves much to be desired, and I was wondering if there's any sort of "So you've just become sysadmin" guides that anyone could recommend.

    Read the article

  • New harddrives failing within weeks.

    - by Jason Kealey
    I've experienced 8 hard disk failures in 3 months and have tried many things to solve the issue permanently but I have failed. I would like to know if you have any advice for me. System was running Win XP on an Asus P5W-DH Deluxe. I have setup a RAID-1 array. I started out with 2 x 500 GB 7200RPM Western Digital drives. One died. I took it out to RMA it. On the same day, the router was fried. Assumed a power surge occurred; connected an older UPS to protect the system. Once I got my hands on an identical disk, I installed it. The RAID array was rebuilt. A few days later, the other one died. Assumed the rebuild caused it to fail. Took it out for RMA. Before the other one arrived, the remaining one died. I then discovered I could re-enable them using the Intel Matrix Storage Manager. I re-enabled both and the system seemed fine for a week, until both died again. I got two new 1.5 TB 7200RPM Seagate drives and re-installed Windows 7. Also replaced the UPS and power supply. They both died again. The voltage on the plug is stable between 120 and 122V as per the UPS. None of the other devices have had any problems (monitors, etc.). At this point, I see two options: a) electrical issue in the house that was, for some reason, not blocked by the UPS. b) something else inside the system causing surges? motherboard? onboard raid controller? Failures happen fairly quickly, between 2 and 14 days after I fix the previous issue. I just gotten a new computer (Core i7) to replace it. If it is stable, I can determine that b) was the problem. If it fries its hard drive again, I can determine that it is an electrical issue in the house. Do you have any other thoughts? Any tools I can run on the drives that failed to get more information about the original SMART event history?

    Read the article

  • Export MS SQL database as *.dbschema

    - by jjczopek
    We have a production database and visual studio 2010 database project. We had to make some changes in database schema. Unfortunately we don't have previous database schema file for production database. Is there a way to export existing database schema as *.dbschema file, preferably from Microsoft SQL Server Management Studio (2008 R2)? This way we could run schema comparison and generate update script.

    Read the article

  • Best intranet solution for a small digital agency?

    - by Rob B
    I work for a small digital agency (10-15 employees). We design and build websites (using php and Django). Are we better off building our own intranet using django or going for a pre built (free or paid for) solution? Key features needed: Project management (basecamp clone) Time reporting Local and external file sharing (dropbox clone)

    Read the article

  • SQL Server Logs: missing date ranges

    - by Jeff
    I need to be able to export SQL Server logs into CSV files, which I can easily do with the export function. However in doing so I've noticed there's a range of dates missing from the SQL Server logs in Management Studio, two months actually. I'm wondering where these logs might be found, and if it's possible to reload them so I can view and then export them. They're strictly for informational purposes, but I do need them. Thanks in advance!

    Read the article

  • I accidentally made my new hard drive (the non system drive) my primary partition

    - by qwerty2
    Hi all, When installing a new hard drive, I accidentally formatted it using 'Disk Management' and set it up as my primary active partition, even though it isn't the system drive. Then, when I restarted my machine, Windows wouldn't boot, citing a missing or corrupt SYSTEM folder. Can anyone help me re-enable the system hard drive as my primary active partition? Any help would be greatly appreciated.

    Read the article

  • Boot win7 from laptop after booting win 8 to go from usb

    - by user1687031
    Just successfully installed and ran windows 8 to go on an external usb hdd. After shutdown and removing the usb hard disk, tried to start my laptop which has a single windows 7 system, but failed to boot and even repair trials didn't succeed. It seems that win 8 affected the strucure of the laptop hard drive which results in corruption in its main win 7 partition. How to fix that and avoid future problems of the same type? thanks.

    Read the article

  • MS Publisher 2003 - hangs when saving to desktop

    - by Chris
    We have a win 7 home prem pc, amd cpu, 8G ram, plenty of free disk space. Whenever user is working in publisher 20003, and tries to save a publisher 2003 document to the desktop, the save as dialog hangs and takes 2-3 minutes to display the desktop save location. I've tested excel 2003, it has no problems immediately displaying the desktop save as location and saving the file.

    Read the article

  • KVM-Guest does not boot: qemudParsePCIDeviceStrs

    - by markus
    I have a Server running Ubuntu 10.10 Server-Edition kvm, and libvirt (both ubuntu-native packages) HDD-Partitioning was done with LVM. Then I created some VMs with Virt-Manager and assigned LVM-Volumes to the VMs. Now the VMs do not boot. Virt-Manager shows a CPU-Usage of 100% for this Guest and the VNC-Connection states Booting from Hard Disk The VM-specific logfiles do not show any abnormality only syslog shows a warning warning : qemudParsePCIDeviceStrs:1422 : Unexpected exit status '1', qemu probably failed What can I do to find the error?

    Read the article

  • Download of ~100MB file from Joomla + Docman component is incomplete (sometimes)

    - by Knix
    Hi, I hope this is right place for this. I am currently running Joomla v1 with a file management component called Docman on a Bluehost server. Some users (particularly with slower download speeds) are experiencing partial file downloads with some of my larger files (~100MB). I would like to keep the Joomla + Docman installation. Is there anything I can do to resolve this issue? I would greatly appreciate any recommendations that you may have. Thanks, Knix

    Read the article

  • Using SSH, transfer webURL to remote machine

    - by AlanTuring
    Hi so i was doing some research in the library so i could use some pictures later on my Desktop computer in my room. I have space on my Lab account which i usually SSH into, and i was wondering if URL's can be directly transferred over to a remote machine and saved on the hard disk. I was thinking something like this: scp http://click.si.edu/images/truncatedurl.jpg /home3/etc.../filename.jpg is this possible? Thanks in advance.

    Read the article

  • 7-Zip compression on multi-core computers

    - by Peter Mortensen
    Does 7-Zip take advantage of multiprocessor or multi-core systems when compressing? For example, would there be a close to 16 times speed-up on a 16 core system assuming no disk or memory bottlenecks? Or is it is limited to 2 threads (2 times speed-up on systems with more than one CPU or core)?

    Read the article

  • iotop for Linux kernel 2.6.18

    - by Lightsauce
    So it has to come to my attention that iotop isn't availalbe for 2.6.18 since it's less than 2.6.20 and requires Python 2.6+. I've done some research and came across this article: http://lserinol.blogspot.com/2009/09/io-usage-per-process-on-linux.html According to this, if these process have io stats in /proc/pid#/io (where pid# is the process #) it's doable regardless of the kernel version. So, in reality, I could upgrade Python to 2.6 and test out iotop. However, my flavor of Linux, CentOS release 5.5 (Final), only supports Python 2.4.3-44.el5 currently. If I were to do uninstall from yum, it doesn't look so pretty. It ends up wanting to uninstall 235 packages, most of which are very important! I read in one place, online (I forget the URL from yesterday), that you can install Python 2.6+ parallel to this one, and have the rpm install for iotop use that. Well, I didn't choose that route. I figured, what the heck, lets write iotop (not copying it, but reverse engineering it without actually looking at it's code/it in use) in bash. I thought it would just grab the /proc/pid#/io file and parse stats. So I wrote a script to grab the top 10 rchar, wchar, read_bytes, and write_bytes by collecting all these stats from all the /proc/pid#/io files, sorting them by each metric, then grabbing the top 10 highest values. The conclusion, the data seems completely useless. Does anybody know any resources for advanced Linux where I can figure out how to take these /proc/pid#/ directories and figure out what the heck they are doing with io on the disk? My main goal is to figure out what exactly is causing high load on my disk. I just know it's on the / partition (/dev/sda2 in this case), and I'm not really sure how to narrow it down without the help of iotop. If I run iostat to grab metrics for 1 minute, every second, the first result it gives me shows a high 'kB_read/s', so that makes me think, it's reading mostly. However, if I watch the update it gives me every second, it's actually just showing values for kB_wrtn/s. This makes me think the initial value iostat gives me is misleading.

    Read the article

  • How to move the files of a replicated database (SQL Server 2008 R2) to a different drive

    - by ileon
    I would appreciate if someone could help me with the following problem: We use two SQL Server 2008 R2 databases under transactional replication: transactional publication with updatable subscriptions. because we run out of disk space we need to move the database files into a new drive. But I don't want to break the replication. What I'm looking for are the required steps that will help me to move the files to the new drive. Thanks

    Read the article

  • ZFS Recover from Faulted Pool State

    - by nickv2002
    I have a six disk ZFS raidz1 pool and had a recent failure requiring a disk replacement. No problem normally, but this time my server hardware died before I could do the replacement (but after and unrelated to the drive failure as far as I can tell). I was able to get another machine from a friend to rebuild the system, but in the process of moving my drives over I had to swap their cables around a bunch until I got the right configuration where the remaining 5 good disks were seen as online. This process seems to have generated some checksum errors for the pool/raidz. I have the 5 remaining drives set up now and a good drive installed and ready to take the place of the drive that died. However, since my pool state is FAULTED I'm unable to do the replacement. root@zfs:~# zpool replace tank 1298243857915644462 /dev/sdb cannot open 'tank': pool is unavailable Is there any way to recover from this error? I would think that having 5 of the 6 drives online would be enough to rebuild the right data, but that doesn't seem to be enough now. Here's the status log of my pool: root@zfs:~# zpool status tank pool: tank state: FAULTED status: One or more devices could not be used because the label is missing or invalid. There are insufficient replicas for the pool to continue functioning. action: Destroy and re-create the pool from a backup source. see: http://zfsonlinux.org/msg/ZFS-8000-5E scan: none requested config: NAME STATE READ WRITE CKSUM tank FAULTED 0 0 1 corrupted data raidz1-0 ONLINE 0 0 8 sdd ONLINE 0 0 0 sdf ONLINE 0 0 0 sdh ONLINE 0 0 0 1298243857915644462 UNAVAIL 0 0 0 was /dev/sdb1 sde ONLINE 0 0 0 sdg ONLINE 0 0 0 Update (10/31): I tried to export and re-import the array a few times over the past week and wasn't successful. First I tried: zpool import -f -R /tank -N -o readonly=on -F tank That produced this error immediately: cannot import 'tank': I/O error Destroy and re-create the pool from a backup source. I added the '-X' option to the above command to try to make it check the transaction log. I let that run for about 48 hours before giving up because it had completely locked up my machine (I was unable to log in locally or via the network). Now I'm trying a simple zpool import tank command and that seems to run for a while with no output. I'll leave it running overnight to see if it outputs anything.

    Read the article

  • "Upgrading" SQL Server 2008 180-day Evaluation to Licenced Standard Edition

    - by alsan
    Hello, I run into the same issue as someone who posted this question on experts-exchange.com (couldn't read the answer though as I don't have an account there): {Quote Begin} I noticed that the 180-day Evaluation version of SQL Server 2008 is the Enterprise version. Is there going to be any problem "upgrading" the Evaluation Enterprise version to a licensed STANDARD version (and how much additional stuff is going to be left inactive on my disk and, more importantly, in my registry, etc. if I do so)? {Quote End} Any advice is appreciated.

    Read the article

  • Why do my VMware Images get so large?

    - by stevebot
    Hi, I have a Centos VMware Image that I have recreated a couple times, and I notice that after a while it gets pretty large. It starts out at 8 GBs when I make it, and a week or two later it is 25GB and then a month later it is a whole 50GB or so. I am not installing anything crazy on it, and my disk usage on the VM is pretty low. Is there an option that could be affecting the size of these VMs?

    Read the article

  • Minimize writes to SSD disks with Windows 7

    - by mfn
    Most people use their SSD as their primary system installation disk with Windows 7. W7 already has a lot of optimizations for SSDs, both in terms of performance and lifetime. Minimizing writes increases the lifetime of SSDs, so post each suggestion as an answer and let others vote on them.

    Read the article

< Previous Page | 452 453 454 455 456 457 458 459 460 461 462 463  | Next Page >