Search Results

Search found 7072 results on 283 pages for 'peopletools 8 50'.

Page 124/283 | < Previous Page | 120 121 122 123 124 125 126 127 128 129 130 131  | Next Page >

  • Missing disk space in Windows XP

    - by Jørn Schou-Rode
    On my mother's Lenovo laptop, Windows XP claims that the hard drive is almost full. According to the properties window, 52.7 out of 55.2 GB is in use: By deleting temp files from Internet Explorer, System Restore, Recycle bin, Windows Update, System Cleanup, I managed to free up about one GB. That's still 50 GB in use, which still is a lot more than I expected. Hence, I gave good old WinDirStat a spin, and here's the output: It might be hard to read here, but the first line says that the total amount of disk space in use on drive C is 24.3 GB. So Windows claims usage of 52.7 GB and WinDirStat can only account for 24.3 GB. Where is the other half of that disk space being used? I hope someone has an answer, or some tricks or tips to do further research. UPDATE: The laptop in question has an SSD hard drive. I am aware that these disk (at least the earlier ones) have a limited life-time. Could the symptoms described be caused by wear and tear on the SSD?

    Read the article

  • Best way to split LAN computers on two WANs (not load balancing)

    - by lpfavreau
    What is the best way to split the computers of a LAN (about 50 computers) to go on two WANs, depending on the computers' IP address or computers' configuration (gateway). I'm not looking for load-balancing, I need to be able to route which computer uses which Internet connection. A solution would be to rewire the old office to split the two distinct groups that need to use two different Internet connections but I'm not sure it's worth it at the moment. I have a pfSense router configured with the LAN (DHCP), WAN1 and WAN2. I also have the IP addresses (and ability to configure) the next router on WAN1 and the one on WAN2. What would you suggest? Routing based on IP addresses ranges? Different gateways? VLANs would be difficult for the same reason as to why rewire is a bit complicated. Thanks!

    Read the article

  • Large file copy from NFS to local disk performance drop

    - by Bernhard
    I'm trying to copy a 200GB file from an NFS mount to a local disk. The local disk is an XFS filesystem on a LVM on top of a RAID 5 system (hardware RAID controler). I'm using rsync to monitor the transfer speed. At the beginning the IO speed is about 200MB/s, stable for the first 18GB. But then the performance drops by a factor of 10-20 and never recovers to the initial rate. Sometimes it reaches about 50-100MB/s but just for a few seconds and then the process seems to hang for a bit. At the same time all file-stat operations on the target filesystem are blocking for a long time (minutes). Also interrupting the copy process blocks for several minutes, a sub-sequent delete of the partly copied file takes also several minutes. Any ideas what could be causing this?

    Read the article

  • Simulating mouse clicks at specific screen coordinates

    - by Matteo Riva
    I would like to be able to bind some keys to mouse clicks done in specific locations. For example: when I press F1 I should get a left mouse click at coordinates 300x350, F2 at 600x350 and so on. Even better if this could be bound to a specific window application so that coordinates could be relative to it instead of the base desktop. Is there a software which allows this? ADDITION: Ok autohotkey is great but I have problems with my particular setup. Quoting my comment below: I'm using it with an old game (championship manager 01/02) which runs in windowed mode (and I have to set win98 compatibility for it to run): I can get the mouse to move but no click goes to the application I have read this FAQ but it didn't help, this is the script I tried: SendMode Play SetKeyDelay, 0, 50, Play F1::Click 42, 191 F2::ControlSend ahk_class main, Click, Championship Manager 01/02 Still no luck: pointer moves but no click goes through.

    Read the article

  • Problem connecting to MySQL with zend

    - by wheresrhys
    I've set up a virtual host on my local machine (windows xp with xampp installed), have installed zend there and am now trying to connect to the MySQLdatabase. I get the following error message: Message: SQLSTATE[HY000] [2003] Can't connect to MySQL server on ''myhost.com'' (10060) All the tips I've found via google haven't helped. I use kaspersky antivirus and have added port 3306 to the allowed list for MySQL and Apache. I've also pinged myhost.com:3306 and get a response (with ip address 67.63.50.51), though localhost:3306 and 127.0.0.1:3306 aren't found. I've checked that other apps can still access MySQL if the virtual host is pointed at them instead and my non zend apps do work. I guess I have to do something further to the port, or maybe force the app to use a different port, but am not sure how to do either. Any help appreciated.

    Read the article

  • Difficulty racking HP ProLiant Gen8 servers

    - by Systemspoet
    We're an all Proliant shop with around 50 servers, mostly DL360s and DL380, from G5's through G7's. We just got our first two G8's in and went to rack them. We were stunned to find out that the new cable management arms protrude almost 1 inch deeper into the rack then previous iterations of the Proliant line. Unfortunately that causes them to occupy the same space as the PDU's in our APC racks. In a non-densely populated section of rack that's no biggie, but in a densely populated section it's impossible to get the cable arm into place without dislodging another machine's power. Has anyone else run into this? Obviously racking machines without cable management arms is not an option. I supposed we could reconfigure our racks but that's a nightmare.

    Read the article

  • Shrink system volume, create new partition, & enlarge current partitions w/o losing data?

    - by studiohack
    My machine has a 300 GB hard drive. I'm planning to reinstall Windows 7 on this machine soon. It has three partitions: a System partition, and two equal data partitions for particular types of data (music, photos, docs, etc). These data partitions are too small for the data they hold, and the system partition is large enough to be shrunk. I want to create four partitions in preparation for the fresh reinstall, a System partition, and three data partitions of varying sizes (40, 50, & 60 GB respectively). My question: Can I shrink the System volume, create another partition, and make the existing two data partitions bigger without losing any data that is currently on the hard drive? Or is it better to format the whole thing, and create partitions after (thus removing the current OS) with a partitioning tool booting from CD? Can I reformat and create partitions with the Windows 7 set up DVD?

    Read the article

  • Online File Sharing that acts just like LAN shared drives, etc.

    - by Dayton Brown
    Hi All, Have a small business client that wants to move their current file share to the web. Specs are as follows, 20 to 30 GB of space, file sizes are normal (nothing more than 50 to 100 mb) 3 users ideal solution would be exact same functionality as windows explorer. CHEAP!!! But not super cheap. I would like to keep it around $20 per user per month. I've explored a bunch of solutions, but they are all a bit on the complicated side. Thanks in advance for the recommendations.

    Read the article

  • Attempting to caue packet loss with netem doesn't work - possibly because of NAT (but delay does work)

    - by tomdee
    I have traffic from a WIFI access point routed via an Ubuntu box. I have two network interfaces which are NATed *filter :INPUT ACCEPT [11:690] :FORWARD ACCEPT [0:0] :OUTPUT ACCEPT [37:6224] -A FORWARD -s 192.168.2.0/24 -i eth1 -o eth0 -m conntrack --ctstate NEW -j ACCEPT -A FORWARD -m conntrack --ctstate RELATED,ESTABLISHED -j ACCEPT COMMIT # Completed on Thu Mar 15 13:37:21 2012 # Generated by iptables-save v1.4.10 on Thu Mar 15 13:37:21 2012 *nat :PREROUTING ACCEPT [0:0] :INPUT ACCEPT [0:0] :OUTPUT ACCEPT [0:0] :POSTROUTING ACCEPT [0:0] -A POSTROUTING -j MASQUERADE COMMIT If I run a ping app on an Android device connected to the WIFI network I can happily ping google. If I use netem to introduce some delay tc qdisc change dev eth0 root netem delay 100ms I can clearly see pings taking longer. If I use netem to introduce some packet loss tc qdisc change dev ifb0 root netem loss 50% then I see no change. Packet loss does work fine for locally generated traffic, just not for traffic coming in over the network that's being NATed. Any ideas how to sort this out?

    Read the article

  • "Bad response to Storage command" when scheduling job with Bacula

    - by Joril
    I have a Bacula setup with 9 clients, and it's working happily. Today I had to add another client, so I went and copied+adapted the existing configuration files from another client, but when I schedule a job for the new client, I get these errors: 20-Mar 17:50 tools-dir JobId 39: Start Backup JobId 39, Job=BackupPresenze2.2012-03-20_17.50.49_04 20-Mar 17:50 tools-dir JobId 39: Using Device "FileStorage" 20-Mar 17:50 presenze2-fd JobId 39: Fatal error: Failed to connect to Storage daemon: bacula.mylan.local:9103 20-Mar 17:50 tools-dir JobId 39: Fatal error: Bad response to Storage command: wanted 2000 OK storage , got 2902 Bad storage From the client I can telnet to bacula.mylan.local:9103 just fine, and jobs for other clients work successfully... What could I check? (Server and client run Ubuntu 10.04, if it's relevant)

    Read the article

  • retransmission of lost TCP segment

    - by tcpip
    What will happen in the following scenario (assume the connection is already established): (stack A) send 10 byte data (stack B) send ACK for 10 byte data (stack B) send 200 byte data (stack B) send 100 byte data (stack B) send 50 byte data (stack A) send ACK for 350 byte data and also send 70 bytes data This segment gets lost and does not reach machine B. (stack B) retransmit 200 byte data (step 3) (stack A) send ACK for 200 byte data and the next expected seq number as the one for 70 bytes data Question: Should 70 bytes data be also transferred with the ACK in step 8? Note that the retransmission timer for step 6 has not expired yet.

    Read the article

  • Why is idle windows VM using so much CPU?

    - by Jeff Shattock
    I have 2 VMs running as guests on a KVM virtualization platform running on Ubuntu 10.04. One VM is an Ubuntu 10.04 system, the other is a Windows 7 system. When both machines are completely logged out, the Linux machine uses 1% CPU, the Windows one 45-50%, according to top. The graphs in virt-manager seem to back this up. Theres nothing installed on the Win7 image that would be running in the background; its as fresh as can be. Why is the Windows VM using so much more than the Linux VM, when both are logged out and idling?

    Read the article

  • YouTube style linking inside HTML5 Video player

    - by Barney
    After a video has played on YouTube, often there will be a follow up link after the video has finished playing in the form of related videos. I'm looking to implement similar into my HTML5 Video. It would look something like this: At the moment, it seems there is nothing in the W3C spec for this (Links inside HTML5 video). Could anyone advise on a method of achieving? I have 50-100 videos on my site on which I would like to implement. Currently I'm thinking a Javascript callback which initiates the Call to action buttons, though I have a feeling this could get complicated...

    Read the article

  • How do I associate server traffic to a domain hosted on that server?

    - by morley
    I have three or four Linux servers, each of which hosts anywhere from 5 to 50 domains. Each domain has its own folder: /www/projectname/web/ Logs go in: /www/projectname/log However, if there's a traffic spike (or, as I see it on my end, a memory usage spike), I'm not sure how to figure out which domain is responsible for the traffic without running tail -f on each of the projects and making an educated guess based on how fast things scroll. There's got to be a better way! There probably is, but I haven't seen it. And the last time I checked, bandwidth monitors only report system-wide load. So if anyone knows how to do this the right way, please let me know. Thanks!

    Read the article

  • supervisord failed to start nagiosapi after reboot, need to run reload manually

    - by Bajingan Keparat
    I have supervisord to start nagiosapi everytime the server starts. The API created a status dump file called status.dat, which will get updated periodically. The following is the conf file that starts the api. [program:nagapi] directory = /home/nagapi user = api command = /bin/bash -c "source /home/nagapi/.virtualenvs/nagapi/bin/activate; /home/nagapi/nagios-api/nagios-api" stdout_logfile = /home/nagapi/supervisor_nagios-api_stdout.log stderr_logfile = /home/nagapi/supervisor_nagios-api_stderr.log Everytime i restart the server, supervisord cannot start the api. stderr.log claims that it cannot find the status.dat file located in /var/cache/nagios3. It seems like the files was not created yet when supervisor tried to run the api the first time. I'm saying this because if i do a supervisorctl reload, everything would reload just fine, and the api would run ok about 50 seconds after the reload command completes. should i change the command option of the conf file to check for

    Read the article

  • web based source control management software [closed]

    - by tom smith
    hi. not sure if this is the right place, but hopefully someone might have thoughts on a solution/vendor. Starting to spec out a project that will require multiple (50-100) developers to be able to manipulate source files/scripts for a large scale project. The idea is to be able to have each app go through a dev/review/test process, where the users can select (or be assigned) the role they're going to have for the given app. I'm looking for web-based, version control, issue tracking, user roles/access, workflow functionality, etc... Ideally, the process will also allow for the reviewed/valid app to then be exported to a separate system for testing on the test server/environment. This can be hosted on our servers, or we can do the colo process. I've checked out Alassian/Collabnet, but any thoughts you can provide would me appreciated as well. thanks

    Read the article

  • Installing linux mint in one partition

    - by sha404
    So, I have a disk with a MBR setup(image below). I've managed to have 50 GB unallocated space for intalling Linux Mint 14. And I want to keep the current windows OS too(but don't want the Mint inside windows). Now I've seen in some tutorials that Linux Mint needs several partitions for bootloader, swap, & home. I don't like to have so many partitions & maybe MBR stuff won't let me create more than one now. So, is it possible to install Linux Mint in one partition only? If it is really impossible than what's the minimum number? & how can I accomplish that? Thanks in advance.

    Read the article

  • timeout with apache & php w/ each virtual host has his own user process

    - by acemtp
    I have 10 unix users in /home/. Each user is for a specific subdomain for example user www in /home/www/public_html is for www.mywebsite. blog in /home/blog/public_html is for blog.mywebsite. 90% is php and 10% ror for the moment i use apache + fastcgi that use SuexecUserGroup to setup the process with the good user. it seems to works but i have a strange behavior where after a few hours/days, the server stop answering (timeout) but the cpu load is still very low (it's a big server), the apache status display lot of "W" Sending Reply states but there's still 50 idle workers so it should be able to answer. in the older server (lot of slower) we add only one user and using mod_php and we never had this issue. is there another way to do that without fastcgi and SuexecUserGroup or do you know what's going wrong?

    Read the article

  • Servers at remote sites vs. centralized servers?

    - by Boden
    Looking for some opinions here. We've got three physical locations and site-to-site VPN between all three. Currently we've got Windows domain controllers at each location, with roughly 50 clients at each. The domains are currently separate, and we're looking at integrating the three sites. Email (Exchange) will be located at the primary site, and RPD is already being used at the secondary branches to hit the app servers also located at the primary site. The bulk of the local user load at the other two sites is just file sharing. What would the main benefits and drawbacks be of replacing the local domain controllers with NAS devices, and only keeping the domain controller(s) at the primary site? (assuming upgrades are coming regardless) Under what circumstances would you choose one setup over the other?

    Read the article

  • how to troubleshoot shutdown problems in windows 7

    - by Ieyasu Sawada
    I have a desktop with windows 7 ultimate installed. Whenever I shut it down from the start button. The text that says that it is logging off and shutting down is present and the monitor blacks out and says no signal. But the lights in the RAM is still on and all the fans in the desktop is still spinning. Why is that? It seems like its not shutting down completely. I have also got disk problems lately. Event id 175 and 50. Which I think is caused by the computer not shutting down completely. How do I troubleshoot this one? Feel free to ask if you need more details, thanks.

    Read the article

  • Putting servers inside a refrigerator? [closed]

    - by Muhammad Jamal Shaikh
    It could be a silly question, but I decided to go for it. I shall be buying 3 servers in the next few weeks to set up a small webfarm at my home. I am told by different people who work in server rooms, that I should keep my servers in an air conditioned room. Which is really expensive, because the temperature here in south asia is between 10 to 50 degrees C. Here comes the funny part: I have an extra fridge in my home, why shouldn't I put the servers inside that fridge? Benefits: I don't have to buy the air conditioner. I don't have to buy the rack mount for the servers. The electricity consumed by the fridge is much much less than AC. Give me your suggestions!

    Read the article

  • choosing the right RAID level for PostgresQL database

    - by Sergey
    Hi, I got an disk array appliance of 8 disks 1T each (UltraStor RS8IP4). It will be used solely by PostgresQL database and I am trying to choose the best RAID level for it. The most priority is for read performance since we operate large data sets (tables, indexes) and we do lots of searches/scans. With the old disks that we have now the most slowdowns happen on SELECTs. Fault tolerance is less important, it can be 1 or 2 disks. Space is the least important factor. Even 1T will be enough. Which RAID level would you recommend in this situation. The current options are 60, 50 and 10, but probably other options can be even better.

    Read the article

  • How to configure auto-logon in Active Directory

    - by Jonas Stensved
    I need to improve our account management (using Active Directory) for a customer support site with 50+ computers. The default "AD"-way is to give each user their own account. This adds up with a lot of administration with adding/disabling/enabling user accounts. To avoid this supervisors have started to use shared "general" accounts like domain\callcenter2 etc and I don't like the idea of everyone knowing and sharing accounts and passwords. Our ideal solution would be to create a group with computers which requires no login by the user. I.e. the users just have to start the computer. Should I configure auto-logon with a single user account like domain\agentAccount? Is there anything else to consider if I use the same account for all users? How do I configure the actual auto-logon with a GPO on the group? Is there a "Microsoft way" without 3rd party plugins? Or is there a better solution?

    Read the article

  • rsync does't work

    - by jspooner
    I set up a new ec2 ubuntu server and I'm unable to use rsync to push a file up. I can ssh to the machine with my keypair. I'm not sure why this looks like it works but finishes in half a second and there is nothing in /home/ubuntu on the server. ? ~ rsync -av -i ~/.ec2/my-keypair ~/Downloads/pushcom.2012-06-26T01-10-04.gz [email protected] building file list ... done >f..t.... gsg-keypair >f..t.... pushwoodcom.2012-06-26T01-10-04.gz sent 3624 bytes received 64 bytes 7376.00 bytes/sec total size is 3392 speedup is 0.92 I've tried 50 different ways of the rsync command but I can't get anything to work. Please help! Thanks

    Read the article

  • Intel Rapid Start Driver Reinstallation doesn't work?

    - by elegantonyx
    I recently had to reformat my entire drive and reinstall Windows 7. Something that bothers me is that when I was downloading the Intel Rapid Start driver (I have a Lenovo Ideapad Y480, and it comes standard with the Rapid Start) and tried to install it, the installer claimed that my system did not meet the requirements in order to install, when it was previously installed by the OEM. My startup times have increased from 21 seconds to log-in screen to 50 seconds. Please help... Specs: Lenovo Ideapad Y480 Intel Core i7-3610 Ivy Bridge (quad-core) Nvidia Geforce 640M LE with 2gb VRAM 8GB RAM 1TB hard drive, 32 GB SSD (OEM installed)

    Read the article

< Previous Page | 120 121 122 123 124 125 126 127 128 129 130 131  | Next Page >