Search Results

Search found 23021 results on 921 pages for 'process monitoring'.

Page 51/921 | < Previous Page | 47 48 49 50 51 52 53 54 55 56 57 58  | Next Page >

  • Prevent Java application from accessing/monitoring/altering clipboard contents

    - by mcstrother
    I'm a student using a service that provides practice questions for standardized tests. The service requires that I access the questions by downloading and running a Java application. If I try to copy anything from any window of my computer (including applications unrelated to the question bank) while the application is running, the copied item is replaced with an obnoxious message asking me to not pirate their copyrighted material. I find this obnoxious, and I also really don't like the idea that any application can slurp up any and all potentially sensitive information that I happen to copy while it's running. Is there are a way to limit the privileges of this application to stop it from doing this? Thank you!

    Read the article

  • Monitoring VPN users on my network

    - by Fen0x
    My company runs filters on contents available for browsing and I have to check if everyone is respecting the browsing policies. Recently I have spotted some users activate OpenVPN on their workstations and i have discovered that they connect on port 443 of an external server and then use a proxy to circumvent the company filters. Is there any method to monitor VPN connections on allowed ports or to avoid the connections?

    Read the article

  • Process that needs a volume starting before volume mounts

    - by user36126
    The destination for incoming CrashPlan backups on my server (11.04) is /media/SeagateBig (SeagateBig is the volume name of my 2TB USB drive). When the server boots, two things happen: 1) SeagateBig auto-mounts and 2) CrashPlan starts. The problem is, that often these two things don't happen in that order. Then I get: Crashplan starts looks for /media/SeagateBig doesn't find it instead of waiting for it, CREATES IT Now it's backing up onto my / filesystem. NOT COOL. Meanwhile, when SeagateBig finally gets around to mounting, it finds that /media/SeagateBig already exists, shrugs, and creates /media/SeagateBig_ as its mount point. What I need is a way for the order to be enforced - where SeagateBig mounts and then and only then the CrashPlan service is started. Unless I learn that CrashPlan can be told to wait for its destination directory, never to create it... which I am also investigating. But the CrashPlanEngine script is installed by the product so I am loath to modify it, as I know I could by having it loop until df greps successfully for "SeagateBig".

    Read the article

  • How does the Ubuntu upgrade process work?

    - by IDWMaster
    How does Ubuntu upgrade seamlessly to a newer distribution, while the operating system is still running? I'm upgrading from 10.10 to 11.04, and I've upgraded several times before, and it's as simple as running update-manager -d and downloading and installing them, then rebooting. How exactly does this work though? How is the upgrade manager able to update the operating system while it is still in use?

    Read the article

  • Create an RSS feed for any webpage, monitor changes and send results by email?

    - by David
    Do you know a Windows 7 software (not an online service) than can create an RSS feed for any webpage ? Here is what I am looking for : I manually select what is a title, what is a link and if needed what is the content that I would like to keep (article content + their related images if they exist). The software create a RSS feed from that. Then it monitors the feed every x hours (or x days at a specified time). If updates are found then send the results to my email address. (It is a little like a combination of the Firefox addon: Autopager (WYSIWYG selection which use xpath) and WebSite-Watcher (RSS software monitoring).

    Read the article

  • Receiving a notification on a Cluster Group Failover

    - by Diego
    Hi all, I have several Windows Cluster set up and I have the need of keeping track of failovers. I'd need to receive a notification of sort whenever a group fails over. I've seen some examples around, but I can't rely on the approach of just sending a message whenever the resources are stopped/restarted as it would generate too many false alarms. In few words, I need to be notified if and only if the group really fails over. I was thinking that probably the best way is monitoring the System Event Log, but, if possible, I'd prefer not having to write a script/program from scratch for this issue. Is there any script/product that already does it? Thanks.

    Read the article

  • Snort monitoring of spanning interface

    - by aHunter
    I have configured a Cisco 3500 switch with a port SPAN and have my snort node (fedora 13) plugged into it. I am running snort as a daemon and have configured a rule to log all tcp traffic but I am only seeing traffic with a destination of the snort node. I know that the SPAN port is working and wanted to know if there is a specific option that I needed to start snort with in order for it to pickup all the traffic? Or is there something that I have missed here? Many thanks.

    Read the article

  • Monitoring remote employees with screenshots and screen sharing?

    - by Lulu
    I'm looking for a way to track and monitor my online employees I hire on Elance, oDesk and such. The tool should be able to: Produce screenshots on set intervals. Provide real-time screen sharing. Preferably track computer usage while user is logged in as "working" + state which task was done in that time. If I have no all-in-one solution, I will go with RealVNC for screen sharing. But I still need a recommendation for the other things. Thanks

    Read the article

  • Linux: How do I use Munin in cPanel to monitor MySQL?

    - by Continuation
    I have a cPanel server running CentOS 5.5. I want to use Munin to monitor MySQL. I went to: Main >> cPanel >> Manage Plugins and selected "Install and keep updated" for Munin and clicked "Save". I got the usual bunch of status updates about the install. At the end I got: Going to read '/home/.cpan/sources/modules/02packages.details.txt.gz' Database was generated on Wed, 02 Mar 2011 18:28:33 GMT ..........................................................................DONE Going to read '/home/.cpan/sources/modules/03modlist.data.gz' Out of memory! Callback called exit. Done Done Done Process Complete As you can see I got an "Out of memory!" message. But after that it said "Process Complete". Was Munin installed? When I went back to "Manage Plugins" it Munin has a check that against "Install and keep updated". So is everything alright? And how do I use Munin now? How do i configure it to monitor MySQL? Where can I see the results? Thanks.

    Read the article

  • Google Search Engine Optimization - The 3 Step Process to Search Engine Domination

    Would you like more targeted traffic to your website? Would you like to know how to get a stream of profitable customers visiting your site for keywords people are actually searching for? Are you overwhelmed by the complexity of SEO? If you answered yes to any of these questions then the easy 3 step Strategy to Google domination is exactly what you are looking for. In this article I reveal the 3 simple tactics for consistent top ten rankings.

    Read the article

  • Step-To-Step Process to Create Coupon Sites

    Creating coupon sites is one of the ways to create money online; in fact it is also one of the most sought after online business these days. It is because of the fact that most individuals these days are always looking for ways to acquire great discounts for the things that they need.

    Read the article

  • upload process times out for different time zones

    - by shilezi
    I have an inhouse(NY) app that cients can upload files to. Usually uploads go pretty quickly for most clients but this particularly cient in the UK always have problems with uploads. not sure if they get any errors but since we log all exceptions, we see this...Error: System.Web.HttpUnhandledException: Exception of type 'System.Web.HttpUnhandledException' was thrown. ---> System.Net.WebException: The operation has timed out at System.Web.Services.Protocols.WebClientProtocol.GetWebResponse(WebRequest request) at System.Web.Services.Protocols.HttpWebClientProtocol.GetWebResponse(WebRequest request) at System.Web.Services.Protocols.SoapHttpClientProtocol.Invoke(String methodName, Object[] parameters) at ... This shouldn't even be an issue because since we noticed they have had this issue before so we bumped up these to maxRequestLength="2097151" executionTimeout="14400". Further investigating this error, i read that it could be a thread timeout since the default is 20minute. A worker process with process id of serving application pool was shutdown due to inactivity. Application Pool timeout configuration was set to 20 minutes. A new worker process will be started when needed. Problem is I am not entirely sure if that really is the case as all other cients mostly North American, have no issues but then their uploads don't seem to go beyond 80mb and the UK client have done a 700mb upload before that i know of. We have tested 750mb before and the whole process took about 15min for upload and processing. Any help on what the real issue here might be? Thanks.

    Read the article

  • Why would the Apache parent process restart silently?

    - by miracle
    I run apache 2.2.9 with mpm prefork on debian lenny. Following http://httpd.apache.org/docs/2.2/mod/prefork.html, I would expect that there is one parent process, running as root and listening as configured, which would start child processes as defined by the Min/Max/etc. directives. I expect the children to be restarted as per MaxRequestsPerChild, but the parent process to stay put with one process id until I restart it manually. Out of a little paranoia, I started monitoring listening ports including process ids. I have a cron job every 20 minutes to run netstat -ap | grep LISTEN and diff the output. Sometimes (about once per day) I see a series of this: 8c8 < tcp6 0 0 [::]:www [::]:* LISTEN 6194/apache2 --- tcp6 0 0 [::]:www [::]:* LISTEN 6607/apache2 10c10 < tcp6 0 0 [::]:https [::]:* LISTEN 6194/apache2 --- tcp6 0 0 [::]:https [::]:* LISTEN 6607/apache2 Over a period of an hour or three, the parent would change its pid at least once every 20 minutes, without any explanation in the log files or any other hint that anything is going wrong. This is not what I expected. What am I missing?

    Read the article

  • Reliance on the compiler

    - by koan
    I've been programming in C and C++ for some time, although I would say I'm far from being expert. For some time I've been using various strategies to develop my code such as unit tests, test driven design, code reviews and so on. When I wrote my first programs in BASIC I typed in long listings before finding they would not run and they were a nightmare to debug. So I learnt to write a small bit and then test it. These days I often find myself repeatedly writing a small bit of code then using the compiler to find all the mistakes. That's OK if it picks up a typo but when you start adjusting the parameters types etc just to make it compile you can screw up the design. It also seems that the compiler is creeping into the design process when it should only be used for checking syntax. There's a danger here of over reliance on the compiler to make my programs better. Are there better strategies than this ? I vaguely remember some time ago an article on a company developing a type of C compiler where an extra header file also specified the prototypes. The idea was that inconsistencies in the API definition would be easier to catch if you had to define it twice in different ways.

    Read the article

  • Nightly backups (and maybe other tasks) causing server alerts

    - by J. Pablo Fernández
    I have two independent alert notification systems for my servers. The server is a virtual machine on Linode and one of the alerts comes from Linode. The other monitoring system we use is New Relic. They are both watching out for IO utilization. Every night I get alerts from both of them as the server is using too much IO. I run quite a few tasks in the middle of the night but the one I confirmed that can cause IO-warnings is running the backups. The backup is done by s3cmd sync. I tried ionice but it still generates the warnings. Getting warnings every night reduces the efficacy of warnings when they happen for real. For Linode I could raise the level at which a warning is issued, but it might mean making the whole thing useless as the level is too high. What would be the proper solution for this?

    Read the article

  • Monitoring/logging a malfunctioning internet connection

    - by Pekka
    I have a mysterious internet connection problem: Every 15-20 minutes, the connection will become very slow, and take 2-3 minutes for anything to load. I've had a technician from the ISP over here to test the hardware, and everything is in pristine condition. They have no other explanation than a configuration error on my machine, a possibility I can exclude 90% because I'm experiencing the same problems with another machine. I will have to monitor the situation now, and I would like to run a program that logs when internet connections become slow. I thought about putting something together using at and wget. Does anybody know of some other tool for this that does this out of the box - maybe something with an adjustable request frequency, logging connection speeds etc.?

    Read the article

  • Understanding Software Development Process

    Software development means, creating software to fulfill the specific needs of the customers. Depending upon their requirements, the software outsourcing firms develop customized software solutions. The client requirement can be an open source software or program for personal use.

    Read the article

  • Level 3 Communications Monitoring

    - by thursdasy
    Does level 3 Communications offer any type of map of current network status? It seems our connection here is getting high latency through their network. Also what is the experience asking them questions about if one of their appliances is malfunctioning or being overloaded because you are receiving bad latency in their network?

    Read the article

  • Long term system health monitor

    - by user30336
    As an experienced user, I sometimes notice that things are not going well with my computer. For example, my backup drive recently started cycling up and down, so I guessed it was probably dying, and replaced it. I detected this with my ears. Windows did not seem to notice or care. There ought to be software that monitors overall system health by keeping track of things like this, so that unusual events or increasing error rates will not be shrugged off. Among other things: disk errors that are recovered, corrupt network packets (at above the baseline expected rate) and crashes of trusted programs are early warnings. Is there any software that tries to use this kind of monitoring to warn of impending trouble?

    Read the article

  • SEO - A Never Ending Process!

    Any business or institution has to have a website to reach out to its customers. This site should also have a good ranking among the search engine results so that it enjoys visits. For this, the site has to be optimized by a good SEO expert. And it has to go on continually for ever and ever.

    Read the article

  • Monitoring used connections on mysql to debug 'too many connections'

    - by Nir
    On LAMP production server I get the 'too many connections' error from MYSQL occasionally, I want to add monitoring to find if the reason is that I exceed the max-connections limit. My question: How can I query from mysql or from mysqladmin the current number of used connections? (I noticed that show status gives total connections and not the currently used ones.)

    Read the article

< Previous Page | 47 48 49 50 51 52 53 54 55 56 57 58  | Next Page >