Search Results

Search found 7129 results on 286 pages for 'battery usage'.

Page 160/286 | < Previous Page | 156 157 158 159 160 161 162 163 164 165 166 167  | Next Page >

  • Adobe Reader seems to use wrong lpr command when printing to CUPS

    - by zebonaut
    Hi, on my debian lenny box, printing via CUPS works fine, e.g. using OpenOffice. When printing from Adobe Reader, however, stuff doesn't work. It appears the problem happens during the first step of the process, when Adobe Reader talks to lpr. I guess this is the reason because I get an error message from Adobe Reader like: The following error occurred while printing... 'usage: lpr [-cdfghlmnpqrstv] [-#num] [-1234 font] [-C class] [-i [numcols]] [-J job] [-Pprinter] [-T title] [-U user] [-wnum] [name ...]' Looks like lpr complains about the way Adobe Reader talks to it, doesn't it? This happens no matter if I print to my laser or to CUPS-PDF. It seems like Adobe Reader sends a command like lpr -P CUPS-PDF -o PageSizeA4 -o blah -o blah Is there anything I can configure to make things work?

    Read the article

  • Port binding conflicts with "switch user" on Windows 7

    - by C-dizzle
    We are using the switch user function within Windows 7 under an active directory network. We have one application in particular that gives us an error: Only one usage of each socket address (protocol/network address/port) is normally permitted. bind Port 10001 Are there any other ports that can only be used at one time that might have an adverse effect on the other user? We try to mentor our users to use the log off function instead of switch user, but that doesn't always happen. As an alternative, is it possible to disable the 'switch user' button on our machines?

    Read the article

  • What causes winsock 10055 errors? How should I troubleshoot?

    - by Tom Kerr
    I'm investigating some issues with winsock 10055 errors on a chain of custom applications (some of which we control, some not) and was hoping to get some advice on techniques to troubleshoot the problem. No buffer space available. An operation on a socket could not be performed because the system lacked sufficient buffer space or because a queue was full. From research, non-paged pool and ports seem to be the only resources which can cause this error. Is there another resource which might cause 10055 errors? Currently, we have perfmon counters setup on the applications and non-paged pool usage looks low in most circumstances. Open TCP connections looks low and I am unaware of another way to monitor ports. Since it only happens in production, we are unable to use more invasive counters. Is there some other tool or procedure you would recommend to diagnose which application is causing the issue?

    Read the article

  • How reliable are Compact Flash drives

    - by bakytn
    For example, USB drives have read/write limits. and they are considered as unreliable and not durable for daily use over many years. My question is, however, about Compact Flash drives. AFAIK they are built almost in the same way as USB flash drives. I was going to use one Compact Flash drive for daily usage and install a lightweight operating system to it (Linux). What kind of limitations do they have? Read/write operations? Just a time? something else? Thank you!

    Read the article

  • When to increase AWS RDS MySQL Server instance to larger CPU/RAM?

    - by rksprst
    I'm wondering at what stage do I need to increase the image for the RDS MySQL server to a larger CPU/RAM instance. The CPU utilization graph is near 0. The Avg Free Memory is around 150MB. The Avg Swap Usage is 420MB. Read Latency is 0-20ms/op it spikes up randomly. Avg write latency is on average 5ms/op but spikes up to 10-20ms/op. Are there some common rules here that I should follow? Thanks!

    Read the article

  • amazon ec2 pricing

    - by Pradyut Bhattacharya
    I m really confused. I was trying to buy hosting at amazon ec2. My site will not be having much of a traffic and i will be installing glassfish and mysql. Usage will be 1gb of ram and around less than 5gb of hardisk and same bandwidth. As mine is a startup, the number of hits per day would be less than 20hits per day, each hit having around 10mins time. How should i calculate the price on the ec2 calculator. Thanks

    Read the article

  • Find which php scripts cause high CPU with php-cgi

    - by Oli
    Background: I maintain a server for a client who has half a dozen Wordpress sites on. They all have the W3 Total Cache plugin installed and eAcellerator is installed (might be APC). All the PHP sites run through a single batch of fastcgi php-cgi processes (it's actually php-fpm but I'm not sure if that makes a difference). Problem: php-cgi's CPU usage is quite high. Not terminally high but high enough to raise an eyebrow. The client wants to add more sites in the future and I want to avoid becoming CPU limited if I can help it. Question: Is there any way I can find the scripts or even just requests that are causing the high CPU. I realise I might not be able to do anything with the results but it would give me a chance.

    Read the article

  • Dell Poweredge 1950 with Perc 5i keeps losing raid config -> "Foreign Configuration Found"

    - by nosage
    The quick and dirty: the machine is a Dell Poweredge 1950, dual xeon quad cores, 8GB of ram, 2 2TB seagate SATAs in (supposed to be raid1) using a Perc 5i raid card. They are hot-swappable with a back-plane. I can build the raid fine and after a little while an install of server 08 r2 will blue screen and restart. When it comes up the raid controller says "Foreign Configuration Found." When I go into the raid configuration panel there is no raid listed but I can import the "foreign config", and the OS will boot up fine, until it blue screens again after a little while. The issue is OS independent. I have tried swapping raid cards, swapping the RAM module on the raid card and swapping the raid battery, all to no avail. Its almost as if there is a loose connection from the raid card to the back plane and both of disks get lost and the raid card drops the config. But it sees the disks fine when it boots back up. The raid card uses a SCSI SAS cable to connect to the back-plane so I guess the next step is to replace that, but... then I might as well replace the back-plane with a SCSI SAS to sata breakout cable, but... then I need a way to power the disks. Sorry for the wall of txt but it would be great to get some thoughts from people who worked with perc raid cards or poweredge servers with this type of issue before. Ironically I want to get this system up and running so I can work on MCITP labs. Thank you for any/all help and feel free to ask questions!

    Read the article

  • grep pattern interpretted differently in 2 different systems with same grep version

    - by Lance Woodson
    We manufacture a linux appliance for data centers, and all are running fedora installed from the same kickstart process. There are different hardware versions, some with IDE hard drives and some SCSI, so the filesystems may be at /dev/sdaN or /dev/hdaN. We have a web interface into these appliances that show disk usage, which is generated using "df | grep /dev/*da". This generally works for both hardware versions, giving an output like follows: /dev/sda2 5952284 3507816 2137228 63% / /dev/sda5 67670876 9128796 55049152 15% /data /dev/sda1 101086 11976 83891 13% /boot However, for one machine, we get the following result from that command: Binary file /dev/sda matches It seems that its grepping files matching /dev/*da for an unknown pattern for some reason, only on this box that is seemingly identical in grep version, packages, kernel, and hardware. I switched the grep pattern to be "/dev/.da" and everything works as expected on this troublesome box, but I hate not knowing why this is happening. Anyone have any ideas? Or perhaps some other tests to try?

    Read the article

  • Bandwidth Control on our Internet Connection

    - by AlamedaDad
    Hi all, I have Covad dual/bonded T1 service in our office coming through a Cisco 1841 and then through a Sonicwall 3060Pro/Enhanced SW firewall. The problem I'm looking for some input on is how to limit the amount of bandwidth any single user/PC can user for downloading a file from the Internet. It's become an issue that when one person happens to download let's say an ~300MB file, normal internet access for the other employees slows to a crawl. I've seen through MRTG that in fact usage of the circuit jumps to the full 3mb for the duration of the download and then drops. Is it possible to control this? I'm not familiar with QOS or the like so I'm not sure. Any help on this would be appreciated. Thanks...Michael

    Read the article

  • Is there a remote file transfer command that preserves nanosecond timestamps?

    - by Denver Gingerich
    I've tried transferring files using scp and rsync on Ubuntu 10.04, but neither of them preserves more than second precision. Here's an example: $ touch test1 $ scp -p test1 localhost:test2 $ ls -l --full-time test* -rw-r--r-- 1 user user 0 2011-01-14 18:46:06.579717282 -0500 test1 -rw-r--r-- 1 user user 0 2011-01-14 18:46:06.000000000 -0500 test2 $ cp -p test1 test2 $ ls -l --full-time test* -rw-r--r-- 1 user user 0 2011-01-14 18:46:06.579717282 -0500 test1 -rw-r--r-- 1 user user 0 2011-01-14 18:46:06.579717282 -0500 test2 $ A straight copy works fine, but scp truncates the timestamp. Are there any tools (preferably similar to scp or rsync in their usage) that do remote file transfers while preserving nanosecond timestamps? I could write a hacky script to do it, but I'd rather not.

    Read the article

  • Ideas for campus Internet Login mechanism?

    - by miCRoSCoPiCeaRthLinG
    Hello, I work at this university and I'm seeking an effective solution for an internet login mechanism. We have a leased-link at our campus, which is shared by both staff & students. All systems (desktops + laptops + handhelds) connect to the internal network via wifi and can then get onto the net. However, a local govt. regulation requires us to keep track of individual internet usage and hence we need a solution (pref. free / opensource) that'll enable us to implement some sort of an authentication mechanism once a user hooks onto the network. One requirement is that the software should be able to authenticate either against LDAP or some other custom user database (MySQL based) or both. Can anyone suggest any such software or mechanism? Most of our servers are Linux based... so something that runs off such a platform will be good. Thanks, m^e

    Read the article

  • Hyper-Threading and Dual-Core, What's the Difference?

    - by Josh Stodola
    In a conversation with the network administator, I mentioned that my machine was a dual-core. He told me it was not. I brought up the task manager, went to the perfomance tab, and showed him that there are two separate CPU usage graphs. I have a quad-core machine at home and it has four graphs. He told there were two graphs on this particular machine because of hyper-threading. I used to have a hyper-thread pentium 4 processor back in the day, but I never fully understood what it meant. So what is the difference between hyper threading and dual-core? And how do you tell which one you have?

    Read the article

  • Where does apache store initialization state for mod_proxy_balancer

    - by khoxsey
    I run apache2 on Ubuntu as a caching load-balancing reverse proxy in front of a group of application servers. I have noticed that the balancer maintains some state for some of the attributes visible in /balancer-manager such as whether an IP is enabled/disabled, load factor, etc. My site has periods of high (and low) usage, and recently as I added a new server to the working group I noticed that the load balancer picked up the new server but had it set to Disabled. I'm curious where that data is stored, and/or how it is initialized.

    Read the article

  • Pause/play AJAX on particular tabs in firefox

    - by bguiz
    Hi, I want to know if there is some method to disable AJAX on particular tabs within Firefox and re-enable them later. My concern is that I have metered bandwidth, and I need to conserve my usage. But I also like to leave several Gmail tabs open in the background. It would be great if I could just hit a "Pause AJAX" button, to stop the contents of that tab from sending or receiving anything, and then later on hit a "Play" button when I want it to start doing its thing again. Any suggestions?

    Read the article

  • Make blink(1) blink a specific color

    - by dimo414
    I'm playing with my new blink(1), and trying to hook it's command line interface into some of my programs, but the docs are a little limited and I'm struggling to figure out how to run even some seemingly simple commands. In particular, the CLI docs say: Usage: blink1-tool <cmd> [options] where <cmd> is one of: .... --blink <numtimes> Blink on/off .... So I try to run blink1-tool --blink 5 and it outputs blink 5 times rgb:0,0,0: and doesn't light up. How do I use the --blink command?

    Read the article

  • Advantages of multiple SQL Server files with a single RAID array

    - by Dr Giles M
    Originally posted on stack overflow, but re-worded. Imagine the scenario : For a database I have RAID arrays R: (MDF) T: (transaction log) and of course shared transparent usage of X: (tempDB). I've been reading around and get the impression that if you are using RAID then adding multiple SQL Server NDF files sitting on R: within a filegroup won't yeild any more improvements. Of course, adding another raid array S: and putting an NDF file on that would. However, being a reasonably savvy software person, it's not unthinkable to hypothesise that, even for smaller MDFs sitting on one RAID array that SQL Server will perform growth and locking operations (for writes) on the MDF, so adding NDFs to the filegroup even if they sat on R: would distribute the locking operations and growth operations allowing more throughput? Or does the time taken to reconstruct the data from distributed filegroups outweigh the benefits of reduced locking? I'm also aware that the behaviour and benefits may be different for tables/indeces/log. Is there a good site that distinguishes the benefits of multiple files when RAID is already in place?

    Read the article

  • Sar data not collected for 10 minutes

    - by Ichorus
    We have a RedHat server whose only job is to run a JBoss server. Monitors said that memory usage spiked (we have the JVM limited to far less than the total memory on the system) and JBoss crashed. We restarted and everything seems ok now. Odd thing is that sar data for 10 min leading up to the crash is simply not there. Load average got up to the 50s. I have seen severely busy systems (350+ Load Average) still collect sar data. Does anyone have any idea what could cause sar to stop collecting data?

    Read the article

  • Graphics and USB devices freezing soon after OS loads

    - by Andrew
    I run Ubuntu/Windows dual boot. Last night I started the upgrade to Ubuntu 12.04, and my computer has not worked since in either Windows or Ubuntu. Here's what I got when I rebooted after the upgrade, and continue to get every time I boot: Gets to GRUB screen OK. Choose Ubuntu - black screen or crazy purple lines. At first I assumed something went wrong with the upgrade (often happens). Choose Windows - works fine, I log in, but soon after that the graphics freeze (sometimes with purple artifacts). The keyboard and mouse (both USB) also lose power at the same instant, and none of the USB ports have power to them. This happens sooner or later every time I boot. Update: the HDD also appears to lose power at the same point. I have tried a live CD, but my computer refuses to boot any CD even after disabling all other boot options in the BIOS. I have disconnected everything except keyboard, mouse, graphics card with one monitor, one RAM sick and HDD; no change. I also took the little battery out to reset CMOS. I am pretty sure no matter how wrong the Ubuntu upgrade went, it wouldn't cause the above symptoms in Windows. So the only explanation I can think of is that a hardware failure occurred at the same time. Some possible causes of this I can think of are: A couple of days before this, I added a third screen (which worked fine). About a week before, my house lost power in a storm (no ill effects over the past few days though). What can I do, other than buy a new motherboard/CPU and hope it works? Unfortunately I don't have another box to swap parts into to test at the moment.

    Read the article

  • Dedicated server not responding. Malicious?

    - by user2801881
    My Dedicated server dies for days on end. As soon as i reboot after about 20 seconds it does again. Then it will just work again and be fine for another week or so. Im convinced its malicious. Not sure what results or readings i can give you so just ask and i will do whatever is needed. netstat (top 20 connections) 7 79.142.88.250 8 120.202.249.19 8 159.226.21.62 8 188.168.38.102 8 202.114.6.37 8 222.62.207.70 9 60.191.35.42 10 112.124.46.186 10 116.228.55.184 10 181.133.218.11 10 222.90.111.146 11 183.136.146.110 12 124.127.51.135 12 92.225.24.24 13 221.176.23.242 15 119.10.115.165 16 17 218.6.224.66 21 116.228.55.217 24 114.112.194.19 top CPU usage seems to add up to about 10% Mailq is empty Thanks in advance

    Read the article

  • Perl script and out of memory errors

    - by Kevin
    We have a midsized server with 48GB of RAM and are attempting to import a list of around 100,000 opt-in email subscribers to a new list management system written in Perl. From my understanding, Perl doesn't have imposed memory limits like PHP, and yet we are continuously getting internal server errors when attempting to do the import. When investigating the error logs, we see that the script ran out of memory. Since perl doesn't have a setting to limit the memory usage (as far as I can tell) why are we getting these errors? I doubt a small import like this is consuming 48GB of ram. We have compromised and split the list into chunks of 10,000, but would like to figure out the root cause for future fixes. This is a CentOS machine with Litespeed as the web server.

    Read the article

  • Apache service hung and incoming requests not accepted

    - by Gnanam
    Hi, My production server is running Apache v2.2.4 with mod_mono v1.2.4 on CentOS release 5.2 (Final). Suddenly, Apache service hung during normal usage time (approx. 1 pm EDT). Traffic is not too high at this time. This is the first time we're noticing this kind of behaviour in our server. I noticed from access log that even subsequent requests are also not received, even though there were incoming requests. I then manually tried to invoke my application call from web browser, it never returned successfully but it was still loading. I found no unusual behaviour/activity in: 1) Apache access_log and error_log 2) No kernel level errors found in /var/log/messages I've no other option but ended up restarting Apache service. Any idea on what would cause Apache to hang and thereby not allowing subsequent incoming requests? How do I debug/diagnose when this happens next time? Experts advice/recommendation on this are highly appreciated.

    Read the article

  • New Computer won't boot up

    - by Reggie
    I brought new components and put a computer together. I installed win7(32bit) and everything was working as it should. However my case came with two USB3 front ports I could not connect because of motherboard did not support USB3, so I swapped out the board for a new motherboad with USB3 connection. Now I put everything back together but it won't bootup. "Memory Ok" light stay on. Motherboad LED is on. There is power, all fans are on but I got no display. There is no activity in back of the systems, none of the peripheral connection in the back are working. I have tried the following but to no avail. Reboot multiple times Unplug power and plug the power back Reset the CMOS battery..let it stays out for 10 minutes or more. Unplug all peripheral device but no luck Reset memory..... nothing happen The only thing I did not do was remove the cpu...but I unplug everything else. Please help, thank you.

    Read the article

  • How to use HTML markup tags inside Bash script

    - by CONtext
    I have crontab and a simple bash script that sends me emails every often containing PHP, NGINX, MYSQL errors from their log files. This is a simplified example. #/home/user/status.sh [email protected] PHP_ERROR=`tail -5 /var/log/php-fpm/error.log` NGINX_ERROR=`tail -5 /var/log/nginx/error.log` MYSQL_ERROR=`tail /var/log/mysqld.log` DISK_SPACE=`df -h` echo " Today's, server report:: ================================== DISK_SPACE: $DISK_SPACE --------------------------------- MEMORY_USAGE: $MEMORY_USAGE ----------------------------------- NGINX ERROR: $NGINX_ERROR ----------------------------------- PHP ERRORS: $PHP_ERROR ------------------------------------ MYSQL_ERRORS: $MYSQL_ERROR ------------------------------------- " | mail -s "Server reports" $EMAIL I know this is a very basic usage, but as you can see, I am trying to separate the errors, but not of the html tags including \n are working. So, my question is, is it possible to use HTML tags to format the text, if not .. then what are the alternatives.

    Read the article

  • How does NTFS compression affect performance?

    - by DragonLord
    I've heard that NTFS compression can reduce performance due to extra CPU usage, but I've read reports that it may actually increase performance because of reduced disk reads. How exactly does NTFS compression affect system performance? Notes: I'm running a laptop with a 5400 RPM hard drive, and many of the things I do on it are I/O bound. The processor is a AMD Phenom II with four cores running at 2.0 GHz. The system is defragmented regularly using UltraDefrag. The workload is mixed read-write, with reads occurring somewhat more often than writes. The files to be compressed include personal documents and selected programs, including several (less demanding) games and Visual Studio (which tends to be I/O bound more often than not).

    Read the article

< Previous Page | 156 157 158 159 160 161 162 163 164 165 166 167  | Next Page >