Search Results

Search found 13151 results on 527 pages for 'performance counters'.

Page 346/527 | < Previous Page | 342 343 344 345 346 347 348 349 350 351 352 353  | Next Page >

  • Windows Experience Index could not be computed

    - by Alexey Ivanov
    I've upgraded recently from Windows Vista to Windows 8. When I try to rate my computer, it accesses DirectX 9 performance, then processes to DirectX 10 tests: And it gets stuck at this point. In 5–10 minutes, it shows error message: The video card is rather old: Mobile Intel 965 Express Chipset Family. I'm pretty sure it does not support DirectX 10. Why does Windows assess it with DirectX 10? And how can I make it skip DirectX 10 tests and get the system rating? The driver was installed automatically by Windows 8 from Windows Update. Version: 8.15.10.2697 Date: 10/01/2012

    Read the article

  • Unmounting a zfs pool while it is shared with sharenfs

    - by Ted W.
    I have a Solaris (open indiana) system which is getting poor disk write performance. In order to enable ZIL in this version of zfs I need to add a line to /etc/system. This will not take affect until I've unmounted and remounted the zpool. The trick is that this spool is shared via nfs to about 200 other servers to host users' home directories. I can guarantee that no users will be accessing the disks during this period of maintenance but I would like to avoid having to issue an unmount for 200 systems in order to unmount the disk on the Solaris box. My question is, with sharenfs, is it necessary to have all systems disconnected before unmounting the filesystem on the host? If it's possible, how do you go about it? I've tried unmounting already, the normal way, and it reports the disk is busy. There is no lsof in Solaris and pfiles (I think that's what it was) does not show anything obviously using the mounts.

    Read the article

  • Unable to Get IIS Response Times with Hyperic Monitoring Tool

    - by jwmajors81
    We have a very large .NET application (vendor) that I am trying to gather performance metrics for using Hyperic. In general I wanted to be able to report response times for each of the components within the application, which include: Web Services ASP.NET Pages MSMQ I am currently unable to successfully monitor the response times for IIS on my windows machines. I have successfully auto-discovered them and I am getting diagnostic information, but am not getting back the response times. After looking online at what other people are seeing, I found that I am missing a tab for the response times which is usually next to the metrics tab. Also, when I look at the configuration screen for IIS I do not see a field which enables me to specify where the log files are located at. Please note that my logs are located at e:\LogFiles\ instead of the usual location because our IT staff doesn't allocate much space for the C: drive. Please note that I have the MSMQ monitors up and running great. Any assistance would be greatly appreciated. Thanks, Jeremy

    Read the article

  • Cheap windows shared hosting service

    - by Elangovan
    Hi I have recently purchased a domain from go daddy, now I am looking for a cheap windows hosting. Following are my requirements Shared Windows hosting Should be cheap in price Should have at least one SQL server Db and one mysql db. Should support atleast asp.net 3.5 and php Will be good if it has support for asp.net mvc (no problem even if it is not available also) Should be able to install third party blog sites. Bandwidth, total space and performance are not very important. Silverlight is also an added advantage (no problem even if it is not available also). There should be no advertisement or banner added by the hosting company in the site. Should have support for subdomains

    Read the article

  • Server 2008 Hard Faults

    - by claw
    Hey all, plase bear with me as I haven't looked at a server in a very long time. The problem I am having is with a Windows 2008 Standard FE Service Pack 2 Intel Xeon X3430 @ 2.40 2.39 GHZ 4 GB Memory 64 Bit There seems to be no problems other than the physical memory peaking at 91%, always with over 100 Hard Faults Per Second. To my understanding hard faults should be fairly rare on a machine with. Are there any logs I can show you? Or investigate myself. The general performance of the machine is ok, i can access SBS2008 and change settings fairly smoothly without hangs etc. However, we connect to the server and do quite a bit of SQL via an application. For a record to retrieve say 20 rows, it can take 20+ seconds. Thanks in advance, Jamie EDIT: What the server is used for: IIS ASP Web Service SQL 2008 List item Exchange unable to upload screenshots due to low reputation - why doesnt my SO work here :)

    Read the article

  • windows server 2008 vs ubuntu 11 [closed]

    - by user472875
    I am working on implementing a custom server application that should be capable of handling a very large volume of traffic. I am aware that this type of question has been asked a lot, but I haven't been able to find a good answer. What I'm really looking for is for a server with given specs which OS will be able to handle a larger traffic faster and more reliably. I do not care about rights management or any other features. I am fairly good with both platforms, and so I would like to pick the OS with better performance on a clean install, and with nothing else running. Thanks in advance.

    Read the article

  • Best choice for off-site backup: dd vs tar

    - by plok
    I have two 1TB single-partition hard disks configured as RAID1, of which I would like to make an off-site backup on a third disk, which I am still to buy. The idea is to store the backup at a relative's house, considerably far away from my place, in the hope that all the information will be safe in the case of a global thermonuclear apocalypse. Of course, this backup would be well encrypted. What I still have to decide is whether I am going to simply tar the entire partition or, instead, use dd to create an image of the disks. Is there any non-trivial difference between these two approaches that I could be overlooking? This off-site backup would be updated no more than two or three times a year, in the best of the cases, so performance should not be a factor to be pondered at all. What, and why, would you use if you were me? dd, tar, or a third option?

    Read the article

  • Worth it to move /var to physical disk vs logical?

    - by Tammer Ibrahim
    Brief question about partition layout. I use an SSD for /, /boot, /usr, & /home partitions. I'd like to move /var to a mechanical disk to minimize writes to the SSD. I'm mainly concerned about maximizing drive life rather than maximizing performance (although I obviously wouldn't want to cripple my server). My mechanical disks consist of two drives sharing LVM, and a third used for nightly rsync backups. I also have a bunch of old 2.5in hard disks lying around. My question is, should I simply create a new LVM volume '/var' on my primary data store, or would it be worth the increased energy consumption (in terms of maximizing the lifetime of the LVMed drives) to install a low volume 2.5in disk to use just for /var? On a more general level my question is about the trade offs of placing OS mounts on the same physical volumes as my data. Thanks for any help!

    Read the article

  • Ubuntu 9.10 x86_32 with PAE causes graphics issues, mainly slowness

    - by widgisoft
    I've tried the 64bit version and found I was constantly hitting a brick wall trying to get 32bit stuff to run; I'd previously used PAE on 9.04 without any issues so figured I'd give it a shot. However, on 9.10 it seems PAE or the process of enabling PAE breaks the nvidia drivers/module somewhat as performance is terrible; I can't even enable desktop effects and there's lots of artifacting on random controls. Disabling the pae image and rebooting fixes the issue however I'm then stuck with "only" 2.7Gb of ram and unable to use the full 8gb that's installed. Is there somthing special I need to do when using pae and nVidia drivers or should I be using 64bit and just figure out how to force run 32bit packages? :-p

    Read the article

  • What does the "Maximum Frequency" number mean in the Windows Resource Monitor?

    - by nhinkle
    In the Windows Resource Monitor's CPU tab, there is a status box and graph for the "Maximum Frequency", right next to the "CPU Usage" values. What does this mean? The value is sometimes over 100% on my system... what could that imply? By looking at CPU-z's real-time report of the processor's clock speed, it seems to be loosely related to what frequency the CPU is running at, which would imply that it means "percent of maximum possible frequency the CPU is running at"; this would be of relevance on systems with SpeedStep and/or TurboBoost technology (or similar). Furthermore, setting the system to "power saving mode" lowers the "maximum frequency" value to around 60%, while setting it to "high performance" mode sets it to around 110%. However, the percentage does not seem to exactly correlate to the CPU speed being shown. What value is this actually representing then?

    Read the article

  • Why does TeamViewer launch a web server?

    - by OverTheRainbow
    Hello I just read this article which I find concerning: "After 90 minutes of troubleshooting, Nash traced the problem to TeamViewer, which he used to remotely administer the client's servers. It turns out the program had opened up its own webserver on the client's machine as soon as Apache went down and in the process made it impossible for the client, a large provider of business software, to restart its proper website." http://www.theregister.co.uk/2009/07/24/teamviewer_snafu/ I started using TeamViewer recently because it's just much better than UltraVNC (runs on both XP and Vista, the server dials out instead of requiring the user to open a port on their router + better performance than UVNC)... but I thought that TeamViewer had both the server and client connect out to TV's web server for data to flow between the two hosts. Why does TV even need to open a web server on the server host? Thank you.

    Read the article

  • VirtualBox: Host OS processor is spiked while guest OS virtual processor is idle?

    - by Greg Mattes
    I'm running Windows XP 32-bit on Windows Vista 64-bit with VirtualBox 3.0.6. Whenever I run the XP VM, Vista (host) reports 100% cpu utilization even though XP (guest) reports between 1-5% cpu utilization. The host box has 2 GB of physical RAM. The guest/vm is configured with 512 MB. The host box has a 64-bit AMD processor. No apps (other than VirtualBox) are running on either host, they're just idling. Any guesses as to why the host processor is spiked? I've enabled various advanced features for the XP guest in the hopes of having better performance: Settings → System → Motherboard: Enable IO APIC Settings → System → Processor: Enable PAE/NX Settings → System → Acceleration: Enable VT-x/AMD-V and Enable Nesting Paging

    Read the article

  • Terminal Services - MS Access Frequently "Not Responding"

    - by jonfhancock
    Exposition: We use a program built in MS Access that I serve via Terminal Services. I just installed a new TS Server with a Quad Core 2.6GHz Xeon, 8GB RAM, and 4 SATA drives in a RAID 0. In installed Server 2008 R2 (64bit obviously). It's only role is TS. The problem: With just a few sessions (under 10), I start getting frequent Not Responding messages in each session. When it happens, the users aren't doing anything particularly taxing, just form navigation and simple insert queries. I can live with some stalls, but it is visually jarring in WS08 because the screen goes gray, and it presents a dialog offering to wait or close with some other options. Questions: Any suggestions for improving performance and reducing hangs? Is it possible to disable the dialog (always wait) and screen graying?

    Read the article

  • Concurrent backups in SQL Server?

    - by Mikey Cee
    We currently have our backups managed by a third party company. There are a bunch of agent jobs created that take full backups (4 times a day) and transaction log backups (4 times an hour). We now want to manage our backups in house, but don't want to disable the third party's jobs until we are sure that we have everything configured correctly internally So I am proposing to have a short period (say, a couple of days) where backups are being taken both by the old and the new system. I am wondering what the ramifications of having these two different systems both manage backups, and the potential pitfalls of having backups taken simultaneously. Is this even supported? If so, and bearing in mind that the system can cope with one backup without any noticeable performance degradation, is it fairly logical to assume that it should be able to cope with two simultaneous backups? Currently the load on the server is fairly light and it rarely struggles. Any advice is appreciated

    Read the article

  • Best way to duplicate databases nightly?

    - by Margaret
    Hey all We just got two new servers, that are running Windows Server 2008. The intent is to make the machines pretty much identical, copying the content of the master to the slave on a nightly basis, so that if anything fails, the second copy can stand in immediately. It doesn't need to be up-to-the-minute mirroring, though I suppose that wouldn't hurt if performance is not affected. The two machines will, amongst other things, each be running an instance of SQL Server 2008. The aim is to duplicate the databases on the master down to the slave on a nightly basis. Unless I'm misunderstanding, the slave databases in mirrored databases require the primary to be present to work correctly; I'm hoping for some solution where we have a second machine that can be up and running with minimal downtime if the first one falls over. Am I misunderstanding mirroring? Is that the best way to do things, or should I use some other mechanism? If so, what?

    Read the article

  • Trying to determine the correct number of XFS allocation groups for postgresql server on Linux

    - by HBlend
    I am running a postgres 8.4.5 server on the linux 2.6.33.7 kernel on an 8 disk raid array with an LSI controller. Most of the tables are around 1GB or less. I know that XFS uses allocation groups (AG) to achieve I/O parallelism. My first question is, does this mean that if two tables are in the same AG, all I/O requests are queued to both of them if either is being read from/written to? If so, I assume I would want to spread my tables across as my allocation groups as possible, correct? Wouldn't this ensure that multiple users querying different tables would get the best performance?

    Read the article

  • Is there any way to stop or pause Windows fake-raid re-synching?

    - by haimg
    I have two physical disks, each one holds two volumes, they are Windows fake-RAID1. E.g.: Disk A: Partition1 Partition2 Disk B: Partition1-Mirror Partition2-Mirror After an unclean reboot, Windows started resynching these as expected. What was not expected, however, is that it started resynching both volumes at the same time. Both disks seek back and forth like crazy, and synching performance is horrible. Question: Is these any option to stop or pause synching of just one volume? I know I can put one of the disks offline, but that will stop synchronization of all volumes on that disk. Note: I have certain reasons why I need to torture myself with Windows RAID. I'm not interested in alternative RAID solutions right now.

    Read the article

  • Snow Leopard: Optimization

    - by Shyam
    Hi, I have bunch of questions: I have a Mac network, which has five Mac's. Right now, they are individually getting software updates. Is there a way to download the patches/security updates in a single place (repository) and point all machines to this location? Personally, I have tools like Monolingual and Onyx, but are there tools you could recommend that affect the performance of the Operating System positively? Tweaks would be nice. Links and pointers, would be really appreciated. I've read about Time machine, is there a way to backup all machines to a network drive using this tool? Thanks!

    Read the article

  • Does Ubuntu 12.04.1 come with everything I need for using virtual servers and are the tools efficient?

    - by orokusaki
    I noticed that Ubuntu 12.04.1 comes with Xen, OpenStack, KVM and other virtualization-related tools. I have used VMWare in the past. If I was to use Xen for visualization, would I see considerable performance lost, since Xen is run on the host OS? Is it even run on the host OS, or is it like VMWare where it's installed below any Linux OS on the machine (embedded, I guess is the word)? Do you have any recommendations on what sort of set up to use with these built-in tools? I have 2 physical servers, side-by-side. Each will need a VM used for Postgres and a VM used as an app server. One will be a failover for the other.

    Read the article

  • What desktop chipsets support AMD's Rapid Virtualization Indexing?

    - by netvope
    Rapid Virtualization Indexing (RVI) is a hardware virtualization feature that can potentially give significant performance improvement. Intel's equivalent is Extended Page Tables. See http://en.wikipedia.org/wiki/Rapid_Virtualization_Indexing for more details. My question is: Does RVI require motherboard or chipset support? If yes, what desktop chipsets support it? I've ordered a Phenom II CPU (which supports RVI) but I haven't bought the motherboard. I'll install XenServer or VMware ESXi (which both support RVI) on the new computer.

    Read the article

  • ATI video cards - unable to use entire monitor (1080p)

    - by Walter White
    Hi all, I have a Dell s2409w, 24" 1080p monitor. With nVidia, I would plug-in the monitor and voila, it automatically knew it was 1080p (1920x1080). I have both a Windows laptop and Ubuntu laptop. Neither is capable of using the fullscreen even though the monitor reports the input is 1080p. I am connecting the monitors via HDMI, is there a 'special' setting I am missing to make this work? Otherwise, I like the performance of my ATI video cards, the drivers seem to be stable and reliable. Thanks, Walter

    Read the article

  • Diagnosing high CPU waiting

    - by Will
    I have a monitoring server that is running icinga/collectd/graphite with about 50 hosts. I have noticed high load/slugging performance on the box. If you take a look at top, you'll see: Cpu(s): 0.6%us, 0.2%sy, 0.0%ni, 7.6%id, 23.4%wa, 0.0%hi, 0.2%si, 0.0%st Notice the HUGE %wa value, which as far as I know means a network or disk bottleneck. ifconfig shows no dropping packets and there's not a ton of bandwidth going on, so that leaves disk issues, right? There's not a lot of disk writing going on either...iotop is reporting we're only writing a little over 1 MB per second and the RAID tool reports everything is A-OK and write caching is enabled. How do I go about trying to figure out how to fix this? UPDATE: iostat -x output is: avg-cpu: %user %nice %system %iowait %steal %idle 0.62 0.10 0.31 9.65 0.00 89.31 Device: rrqm/s wrqm/s r/s w/s rsec/s wsec/s avgrq-sz avgqu-sz await svctm %util sda 0.21 33.34 83.55 16.54 1599.94 399.07 19.97 43.21 416.98 3.71 37.13

    Read the article

  • Is there a Linux-compatible R/C simulator that works with real radios?

    - by Norman Ramsey
    My Dad flies radio-controlled (R/C) aircraft. He used to run a simulator called "RealFlight" which allowed him to connect his actual radio to his computer and fly simulated craft. He learned enough to fly actual planes, but he wants to move up from "trainer" aircraft to higher-performance craft. After some crashes, he'd like to go back to the simulator for a while. The catch: he's given up Windows and is now running Ubuntu. Question: is there an R/C flight simulator that Runs on Ubuntu? Allows you to connect your radio and use it to control the simulator, preferably through a USB port?

    Read the article

  • Apache - Difference between ScriptAlias and WSGIScriptAlias

    - by Jiri Kadlec
    I'm using apache on RHEL Linux server In my /etc/httpd/conf.d/httpd.conf there are two directives: WSGIScriptAlias /apps /var/www/apps <Directory /var/www/apps > Options MultiViews ExecCGI MultiviewsMatch Handlers SetHandler wsgi-script Order allow, deny allow from all </Directory> ScriptAlias /scripts /var/www/scripts <Directory /var/www/scripts > Options MultiViews ExecCGI MultiviewsMatch Handlers SetHandler wsgi-script Order allow, deny allow from all </Directory> What is the difference? I understand that WSGIScriptAlias is restricted for running Python scripts and ScriptAlias also allows running perl scripts. Can I always use ScriptAlias instead of WSGIScriptAlias? Are there any performance advantages of using WSGIScriptAlias instead of ScriptAlias?

    Read the article

  • Using a Raspberry Pi as a VPN?

    - by sudo rm -rf
    So I'm sure many of you have heard of the new Raspberry Pi project. I was looking at messing around with Model B, which has the following relevant specs: Broadcom BCM2835 700MHz ARM1176JZFS processor with FPU and Videocore 4 GPU 256MB RAM Boots from SD card, running the Fedora version of Linux (ARM Version) 10/100 BaseT Ethernet socket USB 2.0 socket So I was curious if it would be possible to create a simple VPN out of this little machine. I do realize that since it's an ARM processor that might mess up quite a few things. Any ideas if this is possible? Just for what it's worth, this would be a personal project so I'm not worried about performance.

    Read the article

< Previous Page | 342 343 344 345 346 347 348 349 350 351 352 353  | Next Page >