Search Results

Search found 9154 results on 367 pages for 'replication jobs job'.

Page 165/367 | < Previous Page | 161 162 163 164 165 166 167 168 169 170 171 172  | Next Page >

  • Forensics on Virtual Private servers [closed]

    - by intiha
    So these days with talks about having hacked machines being used for malware spreading and botnet C&C, the one issue that is not clear to me is what do the law enforcement agencies do once they have identified a server as being a source or controller of attack/APT and that server is a VPS on my cluster/datacenter? Do they take away the entire machine? This option seems to have a lot of collateral damage associated with it, so I am not sure what happens and what are the best practices for system admins for helping law enforcement with its job while keeping our jobs!

    Read the article

  • Reliable file copy (move) process - mostly Unix/Linux

    - by mfinni
    Short story : We have a need for a rock-solid reliable file mover process. We have source directories that are often being written to that we need to move files from. The files come in pairs - a big binary, and a small XML index. We get a CTL file that defines these file bundles. There is a process that operates on the files once they are in the destination directory; that gets rid of them when it's done. Would rsync do the best job, or do we need to get more complex? Long story as follows : We have multiple sources to pull from : one set of directories are on a Windows machine (that does have Cygwin and an SSH daemon), and a whole pile of directories are on a set of SFTP servers (Most of these are also Windows.) Our destinations are a list of directories on AIX servers. We used to use a very reliable Perl script on the Windows/Cygwin machine when it was our only source. However, we're working on getting rid of that machine, and there are other sources now, the SFTP servers, that we cannot presently run our own scripts on. For security reasons, we can't run the copy jobs on our AIX servers - they have no access to the source servers. We currently have a homegrown Java program on a Linux machine that uses SFTP to pull from the various new SFTP source directories, copies to a local tmp directory, verifies that everything is present, then copies that to the AIX machines, and then deletes the files from the source. However, we're finding any number of bugs or poorly-handled error checking. None of us are Java experts, so fixing/improving this may be difficult. Concerns for us are: With a remote source (SFTP), will rsync leave alone any file still being written? Some of these files are large. From reading the docs, it seems like rysnc will be very good about not removing the source until the destination is reliably written. Does anyone have experience confirming or disproving this? Additional info We will be concerned about the ingestion process that operates on the files once they are in the destination directory. We don't want it operating on files while we are in the process of copying them; it waits until the small XML index file is present. Our current copy job are supposed to copy the XML file last. Sometimes the network has problems, sometimes the SFTP source servers crap out on us. Sometimes we typo the config files and a destination directory doesn't exist. We never want to lose a file due to this sort of error. We need good logs If you were presented with this, would you just script up some rsync? Or would you build or buy a tool, and if so, what would it be (or what technologies would it use?) I (and others on my team) are decent with Perl.

    Read the article

  • Apache spawning one process a day, what causes it?

    - by xtrimsky
    I have a PHP server running, with cron jobs etc... And once a day (between midnight and 3am), apache spawns one process that never ends. The server is a virtual server, so in a couple of days it eats up the whole memory. Is there a way to figure out what is crashing ? What php script isn't finishing ? Or what url via apache was triggered ? I have tried looking on the access logs, errors logs, but didn't find anything unusual. Thank you

    Read the article

  • Can I rent exclusive time on a powerful server running linux? [closed]

    - by Mark Borgerding
    My company is involved in a proposal that requires speed estimates of our software on a server with the latest & greatest processors. This is not the first time we've been in this situation. The servers themselves are too expensive to buy a new one every time, so we end up extrapolating from what we have. There are so many variables: processor generation & speed, memory speed, memory channels, cache configurations; it makes extrapolation difficult and error-prone. Is there a business that rents time on the newest servers? At least part of the time we'd need exclusive access to an otherwise quiescent system either via ssh shell access or unattended batch jobs. I am not looking for general cloud computing services. I don't need much time on the server, but it needs to be exclusive. And the server needs to be pretty cutting edge for a solid basis of estimate.

    Read the article

  • Share iTunes Library from Nas across several clients. How in Windows

    - by Mych
    I found this article http://gigaom.com/apple/one-itunes-library-on-multiple-computers/ which describes sharing a single library with multiple clients. Unfortunately this article is for Macs and I use Windows. The article mentions 3 jobs that need to be completed.' A) - pointing all clients to the location of the Library. This is understandable and I can replicate using windows clients. B) Universal Library Set-up. This mentions holing the option key and double click iTunes icon. This is so you can create the index at a specific location C) Point clients to index. Again this mentions Option double click. What is the Windows equivalent to option double click iTunes?

    Read the article

  • Linux: Create files and direcotires but not delete them

    - by Peraz
    I have a process that create directories and files inside a working directory, ex: /workingdir/file1 /workingdir/file2 /workingdir/dir1/file1 /workingdir/dir1/dir2/file1 /workingdir/dir1/file2 I need to avoid deletion/overwrites of created folders/files for that user, but allow subsequent folders/subfolders/files creation. I try permissions, gid, acl with no luck. What is the correct way to do that ? (i can use a cron job if needed)

    Read the article

  • Checking that tasks are executed

    - by homer5439
    I'm not sure how to explain this. Once one starts having dozens or hundres of servers, each running some sort of periodic jobs (mostly from cron), there is a problem of making sure (or as sure as possible) that these tasks are actually ran. I mean, I get an email if a job fails fails, and no mail if it succeeds, but also no mail if it doesn't run for whatever reason. Sure, I could change them and have them send a "successfully ran" email, only to be flooded by mails that most of the time I don't want to see. Basically, I want to be notified only if: a task ran and failed a task didn't run at the expected time. Is there a way to do this?

    Read the article

  • can I use @reboot in cron.d files?

    - by fschwiet
    I want to run a job with cron on reboot as a particular user. I have been able to do this successfully using crontab to write to /var/spool/cron/crontabs/username with something like: @reboot ./run.sh >>~/tracefile 2>&1 However, I want to use /etc/cron.d/filename. Cron jobs in this file require an extra column to indicate what user runs, so I use: @reboot wwwuser ./run.sh >>~/tracefile 2>&1 This doesn't seem to work. Should I be able to use @reboot with a username in a cron.d file?

    Read the article

  • Best server for mailing application [closed]

    - by Cyber Junkie
    My application is similar to a reminder service that reminds users of events that they scheduled. I'm sending emails to users through a PHP script. I'm not sending one email to multiple recipients. Each recipient receives a different message. I plan to use cron jobs every minute and expect the application to send roughly 200 individual emails in 1 hour (for a small user base that may grow). I don't have hosting experience with this type of application. I plan to start on a shared host and move up in the future to vps or dedicated. Most shared hosts that I looked into allow 50-100 emails per hour with delays between mailings. Please kindly inform me what I should look for in web hosts for this kind of application.

    Read the article

  • Shell script for daily disk usage report

    - by Master
    I am doing backups on my local drives. The drives are mounted in /media folder. Now i want to run cron job daily which will tell in table format how much disk is used by folder and how much free space is left on drive It would be good if i can insert that info in database and i can see that info use webpage on locahost ubuntu 10

    Read the article

  • Automated monitoring of a remote system that sends email alerts.

    - by user23105
    I need to monitor a remote system where the only access I have is that I can subscribe to email alerts of completed/failed jobs. I would like a system that can monitor these emails and provide an SMS or other alert when: An email indicates failure. A process that was expected to complete by a given time has not. A process that was expected to complete N minutes after completion of another process has not completed. Are there any existing tools that allow this? I'd consider any option - SaaS, open-source, COTS, as long as I don't have to write it myself! Cheers, Blake

    Read the article

  • How do i network ten branch office with voip, video calling and files sharing.

    - by Oluwalogbon
    Am an IT person, have done some networking job for my organization like Lan and wireless within the area, configure windows server to manage staff account My company has ten branch (In each state) in my country and am giving a task to connect dose branch together, which there will be VOIP, Video calling and sharing of files within the branch. I need someone to help me with this project..what and what did I need to put in place

    Read the article

  • Port forwarding stopped working in my Linksys WRT54G2 Linksys router.

    - by user23490
    How to do it again? I had simply forwarded needed ports (e.g. for counter strike, ftp, http etc)) but now with same system, same OS and same router and settings, it's not working. Tried setting router to "factory defaults" and do everything again. However other jobs are being done like it is connecting to my DSL ISP and I can access Internet easily. Still no success. PS. I tried on both Windows and Ubuntu. On Windows I use it for Counter Strike and others (e.g. host my local FTP server) on Ubuntu.

    Read the article

  • How to ensure dbs are all in sync when restored?

    - by blade
    Hi, In large server environments, how do you handle the issue of backing up SQL Server dbs which may not be in sync with other dbs they rely on? So if I back up DB1 from a server, and it uses another db which is not backed up, doing a restore when the dbs are in differing state could cause problems? It seems like all dependent DBs should be backed up, regardless of size etc, but in my current job (where we're a datacentre company and I'm a .NET Developer), I only backup some of several dependent DBs on a SQL Server instance. Thanks

    Read the article

  • [PHP] Singleton class and using inheritance

    - by Saif Bechan
    I have am working on a web application that makes use of helper classes. These classes hold functions to various operation such as form handling. Sometimes I need these classes at more than one spot in my application, The way I do it now is to make a new Object. I can't pass the variable, this will be too much work. I was wondering of using singleton classes for this. This way I am sure only one instance is running at a time. My question however is when I use this pattern, should I make a singleton class for all the objects, this would b a lot of code replication. Could I instead make a super class of superHelper, which is a singleton class, and then let every helper extend it. Would this sort of set up work, or is there another alternative? And if it works, does someone have any suggestions on how to code such a superHelper class. Thank you guys

    Read the article

  • How to remotely connect to jmx on tomcat using ssh tunnelling and not break ehcache...

    - by Unsavory
    I've followed the instructions in the following link to create my own RMI registry and jmx server on a single port inside tomcat. According to the comments, I need to set -Djava.rmi.server.hostname=localhost. Once I do that, I can indeed connect to my server via jconsole using ssh port forwarding. http://blogs.sun.com/jmxetc/entry/connecting_through_firewall_using_jmx However, I've found it has the very bad side affect of breaking our ehcache replication which uses RMI. It fails complaining that it cannot bootstrap from remote peer localhost. I'm guessing because the peers all have their rmi server hostname set to localhost from setting -Djava.rmi.server.hostname=localhost. Does anyone have a possible workaround to this problem?

    Read the article

< Previous Page | 161 162 163 164 165 166 167 168 169 170 171 172  | Next Page >