Search Results

Search found 19555 results on 783 pages for 'job performance'.

Page 480/783 | < Previous Page | 476 477 478 479 480 481 482 483 484 485 486 487  | Next Page >

  • Linux HA - Best Heartbeat hardware solution

    - by Martino Dino
    Hi all I would ask anyone what is the best layer 2 medium for heartbeat in Linux and how it's best configured. More precisely I've been thinking about a dedicated NIC for that purpose but then i thought that if a switch breaks then i would loose the heartbeat connection for most of the cluster and STONITH 'BUM'!!! Will probably loose my job after :) Distributing the heartbeat onto the main NICs of every node trough a vif sounds reasonable but im not sure if this is the best option (at least the switches are redundant to some extent). Is it possible to use heartbeat over a bonded interface and that sounds reasonable? Do you have any other tip/solution for that issue?

    Read the article

  • Unmounting a zfs pool while it is shared with sharenfs

    - by Ted W.
    I have a Solaris (open indiana) system which is getting poor disk write performance. In order to enable ZIL in this version of zfs I need to add a line to /etc/system. This will not take affect until I've unmounted and remounted the zpool. The trick is that this spool is shared via nfs to about 200 other servers to host users' home directories. I can guarantee that no users will be accessing the disks during this period of maintenance but I would like to avoid having to issue an unmount for 200 systems in order to unmount the disk on the Solaris box. My question is, with sharenfs, is it necessary to have all systems disconnected before unmounting the filesystem on the host? If it's possible, how do you go about it? I've tried unmounting already, the normal way, and it reports the disk is busy. There is no lsof in Solaris and pfiles (I think that's what it was) does not show anything obviously using the mounts.

    Read the article

  • size of extent on LVM2

    - by piotrek
    in LVM1 there was a limit of 65k extends. So size of extent had to been chosen carefully between wasted space on partitions (to big extent) and maximal possible size of logical volume (too small extent). in lvm2 (according to http://docstore.mik.ua/manuals/hp-ux/en/5992-4589/apa.html) the limit is ~16 million extents. so the default size of 4mb gives ~60TB of LV size. so is there any point in making the extent larger than 4-16mb on a desktop? is there any performance degradation or other costs of having big number of extents?

    Read the article

  • Best choice for off-site backup: dd vs tar

    - by plok
    I have two 1TB single-partition hard disks configured as RAID1, of which I would like to make an off-site backup on a third disk, which I am still to buy. The idea is to store the backup at a relative's house, considerably far away from my place, in the hope that all the information will be safe in the case of a global thermonuclear apocalypse. Of course, this backup would be well encrypted. What I still have to decide is whether I am going to simply tar the entire partition or, instead, use dd to create an image of the disks. Is there any non-trivial difference between these two approaches that I could be overlooking? This off-site backup would be updated no more than two or three times a year, in the best of the cases, so performance should not be a factor to be pondered at all. What, and why, would you use if you were me? dd, tar, or a third option?

    Read the article

  • Cheap windows shared hosting service

    - by Elangovan
    Hi I have recently purchased a domain from go daddy, now I am looking for a cheap windows hosting. Following are my requirements Shared Windows hosting Should be cheap in price Should have at least one SQL server Db and one mysql db. Should support atleast asp.net 3.5 and php Will be good if it has support for asp.net mvc (no problem even if it is not available also) Should be able to install third party blog sites. Bandwidth, total space and performance are not very important. Silverlight is also an added advantage (no problem even if it is not available also). There should be no advertisement or banner added by the hosting company in the site. Should have support for subdomains

    Read the article

  • Worth it to move /var to physical disk vs logical?

    - by Tammer Ibrahim
    Brief question about partition layout. I use an SSD for /, /boot, /usr, & /home partitions. I'd like to move /var to a mechanical disk to minimize writes to the SSD. I'm mainly concerned about maximizing drive life rather than maximizing performance (although I obviously wouldn't want to cripple my server). My mechanical disks consist of two drives sharing LVM, and a third used for nightly rsync backups. I also have a bunch of old 2.5in hard disks lying around. My question is, should I simply create a new LVM volume '/var' on my primary data store, or would it be worth the increased energy consumption (in terms of maximizing the lifetime of the LVMed drives) to install a low volume 2.5in disk to use just for /var? On a more general level my question is about the trade offs of placing OS mounts on the same physical volumes as my data. Thanks for any help!

    Read the article

  • Experience with HAVP

    - by Kyle Brandt
    I have employees that have to search sketchy virus-ridden websites as part of their job. They all have XP and Symantec AV installed but they still get hit fairly often. Rather then trying all sorts of desktop level solutions, I was thinking about maybe some sort of Proxy in addition to symantec for when they browse these sites? Anyone have experience with HAVP? Am I going about this the wrong way? If this is the right way, and free alternatives to HAVP?

    Read the article

  • windows server 2008 vs ubuntu 11 [closed]

    - by user472875
    I am working on implementing a custom server application that should be capable of handling a very large volume of traffic. I am aware that this type of question has been asked a lot, but I haven't been able to find a good answer. What I'm really looking for is for a server with given specs which OS will be able to handle a larger traffic faster and more reliably. I do not care about rights management or any other features. I am fairly good with both platforms, and so I would like to pick the OS with better performance on a clean install, and with nothing else running. Thanks in advance.

    Read the article

  • Server 2008 Hard Faults

    - by claw
    Hey all, plase bear with me as I haven't looked at a server in a very long time. The problem I am having is with a Windows 2008 Standard FE Service Pack 2 Intel Xeon X3430 @ 2.40 2.39 GHZ 4 GB Memory 64 Bit There seems to be no problems other than the physical memory peaking at 91%, always with over 100 Hard Faults Per Second. To my understanding hard faults should be fairly rare on a machine with. Are there any logs I can show you? Or investigate myself. The general performance of the machine is ok, i can access SBS2008 and change settings fairly smoothly without hangs etc. However, we connect to the server and do quite a bit of SQL via an application. For a record to retrieve say 20 rows, it can take 20+ seconds. Thanks in advance, Jamie EDIT: What the server is used for: IIS ASP Web Service SQL 2008 List item Exchange unable to upload screenshots due to low reputation - why doesnt my SO work here :)

    Read the article

  • Trying to determine the correct number of XFS allocation groups for postgresql server on Linux

    - by HBlend
    I am running a postgres 8.4.5 server on the linux 2.6.33.7 kernel on an 8 disk raid array with an LSI controller. Most of the tables are around 1GB or less. I know that XFS uses allocation groups (AG) to achieve I/O parallelism. My first question is, does this mean that if two tables are in the same AG, all I/O requests are queued to both of them if either is being read from/written to? If so, I assume I would want to spread my tables across as my allocation groups as possible, correct? Wouldn't this ensure that multiple users querying different tables would get the best performance?

    Read the article

  • Ubuntu 9.10 x86_32 with PAE causes graphics issues, mainly slowness

    - by widgisoft
    I've tried the 64bit version and found I was constantly hitting a brick wall trying to get 32bit stuff to run; I'd previously used PAE on 9.04 without any issues so figured I'd give it a shot. However, on 9.10 it seems PAE or the process of enabling PAE breaks the nvidia drivers/module somewhat as performance is terrible; I can't even enable desktop effects and there's lots of artifacting on random controls. Disabling the pae image and rebooting fixes the issue however I'm then stuck with "only" 2.7Gb of ram and unable to use the full 8gb that's installed. Is there somthing special I need to do when using pae and nVidia drivers or should I be using 64bit and just figure out how to force run 32bit packages? :-p

    Read the article

  • Why does TeamViewer launch a web server?

    - by OverTheRainbow
    Hello I just read this article which I find concerning: "After 90 minutes of troubleshooting, Nash traced the problem to TeamViewer, which he used to remotely administer the client's servers. It turns out the program had opened up its own webserver on the client's machine as soon as Apache went down and in the process made it impossible for the client, a large provider of business software, to restart its proper website." http://www.theregister.co.uk/2009/07/24/teamviewer_snafu/ I started using TeamViewer recently because it's just much better than UltraVNC (runs on both XP and Vista, the server dials out instead of requiring the user to open a port on their router + better performance than UVNC)... but I thought that TeamViewer had both the server and client connect out to TV's web server for data to flow between the two hosts. Why does TV even need to open a web server on the server host? Thank you.

    Read the article

  • What does the "Maximum Frequency" number mean in the Windows Resource Monitor?

    - by nhinkle
    In the Windows Resource Monitor's CPU tab, there is a status box and graph for the "Maximum Frequency", right next to the "CPU Usage" values. What does this mean? The value is sometimes over 100% on my system... what could that imply? By looking at CPU-z's real-time report of the processor's clock speed, it seems to be loosely related to what frequency the CPU is running at, which would imply that it means "percent of maximum possible frequency the CPU is running at"; this would be of relevance on systems with SpeedStep and/or TurboBoost technology (or similar). Furthermore, setting the system to "power saving mode" lowers the "maximum frequency" value to around 60%, while setting it to "high performance" mode sets it to around 110%. However, the percentage does not seem to exactly correlate to the CPU speed being shown. What value is this actually representing then?

    Read the article

  • Concurrent backups in SQL Server?

    - by Mikey Cee
    We currently have our backups managed by a third party company. There are a bunch of agent jobs created that take full backups (4 times a day) and transaction log backups (4 times an hour). We now want to manage our backups in house, but don't want to disable the third party's jobs until we are sure that we have everything configured correctly internally So I am proposing to have a short period (say, a couple of days) where backups are being taken both by the old and the new system. I am wondering what the ramifications of having these two different systems both manage backups, and the potential pitfalls of having backups taken simultaneously. Is this even supported? If so, and bearing in mind that the system can cope with one backup without any noticeable performance degradation, is it fairly logical to assume that it should be able to cope with two simultaneous backups? Currently the load on the server is fairly light and it rarely struggles. Any advice is appreciated

    Read the article

  • VirtualBox: Host OS processor is spiked while guest OS virtual processor is idle?

    - by Greg Mattes
    I'm running Windows XP 32-bit on Windows Vista 64-bit with VirtualBox 3.0.6. Whenever I run the XP VM, Vista (host) reports 100% cpu utilization even though XP (guest) reports between 1-5% cpu utilization. The host box has 2 GB of physical RAM. The guest/vm is configured with 512 MB. The host box has a 64-bit AMD processor. No apps (other than VirtualBox) are running on either host, they're just idling. Any guesses as to why the host processor is spiked? I've enabled various advanced features for the XP guest in the hopes of having better performance: Settings → System → Motherboard: Enable IO APIC Settings → System → Processor: Enable PAE/NX Settings → System → Acceleration: Enable VT-x/AMD-V and Enable Nesting Paging

    Read the article

  • wireless access point

    - by Warren Bullock III
    I'm hoping to get some suggestions for possible Wireless Access point/router models which will allow us to have two separate networks. We run an internal network on 10.x.x.x IP range where we have shares and other network resources for which we would like to have our regular users access. However, we would also like to offer a separate wireless network for guests which ideally would be on 192.168.x.x and these users would not be able to see any of the resources sitting on the 10.x.x.x network. Anyone have any recommendations on single devices that might be able to get the job done? I was looking at the Linksys E4200 and it seems to support what I'm looking to do... any others? Thanks in advance for any suggestions.

    Read the article

  • Terminal Services - MS Access Frequently "Not Responding"

    - by jonfhancock
    Exposition: We use a program built in MS Access that I serve via Terminal Services. I just installed a new TS Server with a Quad Core 2.6GHz Xeon, 8GB RAM, and 4 SATA drives in a RAID 0. In installed Server 2008 R2 (64bit obviously). It's only role is TS. The problem: With just a few sessions (under 10), I start getting frequent Not Responding messages in each session. When it happens, the users aren't doing anything particularly taxing, just form navigation and simple insert queries. I can live with some stalls, but it is visually jarring in WS08 because the screen goes gray, and it presents a dialog offering to wait or close with some other options. Questions: Any suggestions for improving performance and reducing hangs? Is it possible to disable the dialog (always wait) and screen graying?

    Read the article

  • How do I open a 60M png image on OSX

    - by Topener
    Alright, so I've been looking around on this site on how to open a big PNG image. The question I found was about a 10M png. Xee apparently did the job. So, I downloaded Xee for my 60M file, but it crashes. So does iPhoto, Pixelmator and previewer. In the Pixelmator and Xee case, I actually had to kill the computer, and restart it. It crashed so 'hard' I couldn't get it to respond again. How do I open this file? (and zoom it) Specs: newly acquired macbook pro: 4gb memory, 2.3ghz i7 58M png image approx: 15000x30000 pixels

    Read the article

  • Best way to duplicate databases nightly?

    - by Margaret
    Hey all We just got two new servers, that are running Windows Server 2008. The intent is to make the machines pretty much identical, copying the content of the master to the slave on a nightly basis, so that if anything fails, the second copy can stand in immediately. It doesn't need to be up-to-the-minute mirroring, though I suppose that wouldn't hurt if performance is not affected. The two machines will, amongst other things, each be running an instance of SQL Server 2008. The aim is to duplicate the databases on the master down to the slave on a nightly basis. Unless I'm misunderstanding, the slave databases in mirrored databases require the primary to be present to work correctly; I'm hoping for some solution where we have a second machine that can be up and running with minimal downtime if the first one falls over. Am I misunderstanding mirroring? Is that the best way to do things, or should I use some other mechanism? If so, what?

    Read the article

  • How many cron jobs are too many?

    - by guitar-
    I have a couple of cron jobs for basic maintenance which aren't very resource-intensive. I also have custom task scheduling (which is just calling a .php file and passing information via GET, ie: cronjob.php?param1=param ...). These can add up pretty quickly. These just call system commands and run external programs (Nmap is one of them). They usually don't take long either. Anyway, can anyone tell me, roughly what point is too many? I know it's hard to say since it depends on what job is being run and how often, but at what point does the crontab program start "struggling"? Anyone have any idea? Thanks.

    Read the article

  • Procmail Mailing List (With Access Control)

    - by bradlis7
    This seems like it should be fairly easy to do, but I've run into a few problems. I've added a cron job to parse all users whose UID is greater than 5000: * * * * * root /usr/bin/test /etc/passwd -nt ~allusers/.forward \ && /bin/egrep '([5-9]|[0-9]{2})[0-9]{3}' /etc/passwd | /bin/grep -v 65534 \ | /bin/cut -d ':' -f 1 > ~allusers/.forward Then I created a .procmailrc file: VERBOSE=yes LOGFILE=/var/log/procmailrc #Allow only certain users to send :0 * ^From.*[email protected].* {} :0E /dev/null But, the .forward file is processed before it even gets to procmail, evidently. If I moved the .forward file to another filename, can I use it in procmail to send an email to the users in this file?

    Read the article

  • Cabling: What to keep in stock?

    - by pehrs
    I have worked a few different places, each one with a different solution, so I would like to hear your suggestions and ideas. I am looking at a situation with multiple server-rooms. We have a mixture of copper (both ethernet and telephone), multi-mode fibre and single-mode fibre. We have all types of connectors: RJ-11, RJ-45, LC, SC, FC, SMA and several I probably forgot about. We have a lot of people working in the area, and keeping track of cables is turning into a full time job. So, here are a the questions: What types and lengths of cable do you keep in stock on site? Do you make your own or buy pre-made? Adapters? Dampeners? How do you manage cable inventory? How do you label the cables? Any other tricks to stop this from driving me crazy?

    Read the article

  • Is there any way to stop or pause Windows fake-raid re-synching?

    - by haimg
    I have two physical disks, each one holds two volumes, they are Windows fake-RAID1. E.g.: Disk A: Partition1 Partition2 Disk B: Partition1-Mirror Partition2-Mirror After an unclean reboot, Windows started resynching these as expected. What was not expected, however, is that it started resynching both volumes at the same time. Both disks seek back and forth like crazy, and synching performance is horrible. Question: Is these any option to stop or pause synching of just one volume? I know I can put one of the disks offline, but that will stop synchronization of all volumes on that disk. Note: I have certain reasons why I need to torture myself with Windows RAID. I'm not interested in alternative RAID solutions right now.

    Read the article

  • What desktop chipsets support AMD's Rapid Virtualization Indexing?

    - by netvope
    Rapid Virtualization Indexing (RVI) is a hardware virtualization feature that can potentially give significant performance improvement. Intel's equivalent is Extended Page Tables. See http://en.wikipedia.org/wiki/Rapid_Virtualization_Indexing for more details. My question is: Does RVI require motherboard or chipset support? If yes, what desktop chipsets support it? I've ordered a Phenom II CPU (which supports RVI) but I haven't bought the motherboard. I'll install XenServer or VMware ESXi (which both support RVI) on the new computer.

    Read the article

  • ATI video cards - unable to use entire monitor (1080p)

    - by Walter White
    Hi all, I have a Dell s2409w, 24" 1080p monitor. With nVidia, I would plug-in the monitor and voila, it automatically knew it was 1080p (1920x1080). I have both a Windows laptop and Ubuntu laptop. Neither is capable of using the fullscreen even though the monitor reports the input is 1080p. I am connecting the monitors via HDMI, is there a 'special' setting I am missing to make this work? Otherwise, I like the performance of my ATI video cards, the drivers seem to be stable and reliable. Thanks, Walter

    Read the article

< Previous Page | 476 477 478 479 480 481 482 483 484 485 486 487  | Next Page >