Search Results

Search found 3841 results on 154 pages for 'daily deals india'.

Page 104/154 | < Previous Page | 100 101 102 103 104 105 106 107 108 109 110 111  | Next Page >

  • apache chokes after 300 connections

    - by john titus
    We have an apache webserver in front of Tomcat hosted on EC2, instance type is extra large with 34GB memory. Our application deals with lot of external webservices and we have a very lousy external webservice which takes almost 300 seconds to respond to requests during peak hours. During peak hours the server chokes at just about 300 httpd processes. ps -ef | grep httpd | wc -l =300 I have googled and found numerous suggestions but nothing seems to work.. following are some configuration i have done which are directly taken from online resources. I have increased the limits of max connection and max clients in both apache and tomcat. here are the configuration details: //apache <IfModule prefork.c> StartServers 100 MinSpareServers 10 MaxSpareServers 10 ServerLimit 50000 MaxClients 50000 MaxRequestsPerChild 2000 </IfModule> //tomcat <Connector port="8080" protocol="org.apache.coyote.http11.Http11NioProtocol" connectionTimeout="600000" redirectPort="8443" enableLookups="false" maxThreads="1500" compressableMimeType="text/html,text/xml,text/plain,text/css,application/x-javascript,text/vnd.wap.wml,text/vnd.wap.wmlscript,application/xhtml+xml,application/xml-dtd,application/xslt+xml" compression="on"/> //Sysctl.conf net.ipv4.tcp_tw_reuse=1 net.ipv4.tcp_tw_recycle=1 fs.file-max = 5049800 vm.min_free_kbytes = 204800 vm.page-cluster = 20 vm.swappiness = 90 net.ipv4.tcp_rfc1337=1 net.ipv4.tcp_max_orphans = 65536 net.ipv4.ip_local_port_range = 5000 65000 net.core.somaxconn = 1024 I have been trying numerous suggestions but in vain.. how to fix this? I'm sure m2xlarge server should serve more requests than 300, probably i might be going wrong with my configuration.. The server chokes only during peak hours and when there are 300 concurrent requests waiting for the [300 second delayed] webservice to respond. Please help..

    Read the article

  • SBS Report displaying critical error events but nothing in Event Viewer

    - by evesirim
    We have a Windows Server server set up running SBS 2008 on a 64-bit architecture. When I run my daily SBS Console reports I am being told there is a critical error on the server with Event ID 56 & Source "TermDD". This is all fairly regular behaviour and I have plenty of sites I would normally troubleshoot errors on but the strange thing is there are no errors appearing in the Windows Event Viewer & I can find no help on why this might be... Has anyone experienced this before / know what might be the cause? Many thanks

    Read the article

  • Duplication of Windows 7 Backup

    - by Steven Pickles
    I use the built in backup utility for Windows 7 because it's automated and flexible enough to allow me to schedule a daily shadow copy backup of particular files and folders directly to a separate internal RAID 0 array (2 x 1TB). It's also lightweight and stays out of the way. For off-site backup purposes, each week I copy the contents of the internal backup from the RAID 0 array to an external 1 TB drive. I then store move this drive to a different building. The copy from the internal backup to the external backup typically works like this: mount and erase contents of external drive highlight "file" on internal drive, hit CTRL+C CTRL+V on root directory of external drive Is there a better way to synchronize? Microsoft's SyncToy application does a pitiful job, and often leaves the folders not truly synchronized... which completely defeats the ability to use the backup's restore feature.

    Read the article

  • Shortcut for enabling / disabling Greasemonkey (or specific script)

    - by ldigas
    talking about firefox here ... I don't use Greasemonkey on any other browser, but if you know, do add info for those as well I use Greasemonkey daily ... having some 20 scripts loaded all the time which save me a ton of grief. But, some of them I sometimes don't need ... few examples: - I use Google Image Status Reporter & Direct Images (links you directly to image file) ... but sometimes I want to go to the page where the image is ... - GReader Minimalist Style ... until I actually need to check the trends and some stuff it hides - there are other examples but these two first sprang to mind, since I just were thinking about that ... To put the long story short ... sometimes I would like a shortcut to disable Greasemonkey, so I don't have to go into the menubar and so on (which I also have half hidden for space) ... anyone knows of any, or how one could create one ?

    Read the article

  • Is there a time machine equivalent for windows that can back up network files?

    - by Jim Thio
    This question is similar to Does an equivalent of Time Machine exist for Windows?, with one difference: The files I want to back up are on a network drive. The computer on that network drive is running Windows XP. I want to back up data on Windows 7. How would I do so? I'd like something similar to Mac OS X' time machine. So copy of data every hour, day, week. Then thinning out, data gets deleted automatically as time goes by. For example, the data for last day is kept as hourly snapshots. For last week, as daily snapshots every day. And for last month as weekly snapshots. How can I achieve this?

    Read the article

  • Only allow root to change filesystem

    - by Uejji
    The VPS I manage uses a simple hard link rsync archive daily backup system saved to a loop file. This is great, because each backup only takes up as much space as what has changed each day, and all user/group permissions are kept. I would like to give users direct access to their home directories in each backup, but I'm worried about intentional or accidental backup data destruction, as how it stands now users can actually change, destroy or add to backed up data they originally owned. I've been looking for a way to mount this filesystem similar to an ro mount option, but something that would still allow rw access to root, but I've had absolutely no luck. In other words, I want users to be able to view and copy their backed up data without actually being able to change it, and have that data maintain the original permissions. I've got no real preferences as far as filesystem, as long as it's a standard unix filesystem that can preserve permissions, support hard links and deny write access to users without actually stripping the w permission from everything.

    Read the article

  • Dual boot centOS and Win7

    - by user1855965
    I posted this on stackoverflow, but it looks like superuser would be more appropriate. I have a CentOS 5 machine that runs Windows 7 as a dual boot. CentOS is the main OS and each OS is set up in a specific hard drive. This was set up before I joined the company and I don't really have need to run Windows now. My question is: can I, from CentOS, reformat the Windows HD, change GRUB settings and get the HD to be available on CentOS? Happy to provide more info if this helps. Many thanks for your help and apologies if this is a very simple issue... I don't want to blindly test things on this machine as it is used on a daily basis by several users.

    Read the article

  • who delete my files?

    - by akalter
    I have some linux server. on two of our server we have mysql. we have daily backup on both machine. but the script different. i saw both scripts. on one of them i saw the "delete older files" algorithm, but in the other this is happening but not from the script. i trying to know who dletes my files, because of that i want to use same script on both machine because of that in the script with the deletion i also copy the files to the another server, and i want to do that in both servers. Who have an idea who delete my older backups? Thank you!

    Read the article

  • Emacs org-mode: how to avoid duplicate lines in agenda, when items is scheduled AND has deadline

    - by Martin
    Many of my TODO items in Emacs org-mode have a DEADLINE defined in the future (e. g. Friday) and are at the same time SCHEDULED today so that I already know I have to start working on this task. Then, this task will appear twice in my agenda. That's not nice but not necessarily a problem yet, but if then the task has assigned a time estimate for its duration and I go to column view with C-C C-X C-C to see how much time my tasks today will need, the time estimate for this task is counted twice, so e. g. if the time effort estimate is 2 hours, I'll have 4 hours in my daily agenda, as the item appears as well as scheduled today (or in the past) as also with its deadline in 3 days. How can I avoid counting an item twice?

    Read the article

  • Disk fragmentation when dealing with many small files

    - by Zorlack
    On a daily basis we generate about 3.4 Million small jpeg files. We also delete about 3.4 Million 90 day old images. To date, we've dealt with this content by storing the images in a hierarchical manner. The heriarchy is something like this: /Year/Month/Day/Source/ This heirarchy allows us to effectively delete days worth of content across all sources. The files are stored on a Windows 2003 server connected to a 14 disk SATA RAID6. We've started having significant performance issues when writing-to and reading-from the disks. This may be due to the performance of the hardware, but I suspect that disk fragmentation bay be a culprit at well. Some people have recommended storing the data in a database, but I've been hesitant to do this. An other thought was to use some sort of container file, like a VHD or something. Does anyone have any advice for mitigating this kind of fragmentation?

    Read the article

  • who deleted my files?

    - by akalter
    I have some linux servers. On two of our server we have MySQL. We have daily backup on both machine. But the scripts are different. I saw both scripts. On one of them I saw the "delete older files" algorithm, but in the other this is happening but not from the script. I am trying to discover who deletes my files, because of that I want to use same script on both machine because of that in the script with the deletion I also copy the files to the another server, and I want to do that in both servers. Who have an idea who deleted my older backups? Thank you!

    Read the article

  • rsync (or robocopy) from multiple computers - what happens?

    - by TheCleaner
    If a Linux server has two different rsync jobs nightly for the same folder to two different destinations, do both destinations end up with the same end set of files? Or does the first job run, and set something on the source folder/files that would cause the 2nd rsync job to not realize the daily changes/updates to the source? Same for a Windows environment using something like robocopy, or even a "differential" backup using BUE or similar. Does each "sync" compare the destination to the source and update the destination regardless of if it is synced multiple times to different destinations?

    Read the article

  • Constructor and Destructor of a singleton object called twice

    - by Bikram990
    I'm facing a problem in singleton object in c++. Here is the explanation: Problem info: I have a 4 shared libraries (say libA.so, libB.so, libC.so, libD.so) and 2 executable binary files each using one another shared library( say libE.so) which deals with files. The purpose of libE.so is to write data into a file and if the executable restarts or size of file exceeds a certain limit it is zipped and a new file is created with time stamp in name. It is using singleton object. It exports a handler class for getting and using singleton. Compressing only happens in the above said two cases. The user/loader executable can specify the starting name of file only no other control is provided by handler class. libA.so, libB.so, libC.so and libD.so have almost same behavior. They all have a class and declare and object of an handler which gets the instance of the singleton in libE.so and uses it for further purpose. All these libraries are linked to two executable binary files. If only one of the two executable runs then its fine, But if both executable runs one after other then the file of the first started executable gets compressed. Debug info: The constructor and destructor of the singleton object is called twice.(for each executable) The object of singleton is a static object and never deleted. The executable is not able to exit/return gives: glibc detected * (exe1 or exe2): double free or corruption (!prev): some_addr * Running with binaries valgrind gives that the above error is due to the destructor of the singleton object. Thanks

    Read the article

  • Upgrading Fedora to Current Version - Dependent Applications Concerns

    - by nicorellius
    Is it possible to upgrade to Fedora 11 and still keep the server functional (it's running Fedora Core release 6 now, with MySQL 5.0, PHP 5.1 and Apache 2.2)? I have a Customer Relationship Management system on there and it is required on a daily basis. I am trying to figure out if I should upgrade (and whether that will cause huge problems) or if I should start over with a new OS (like Ubuntu) and install the CRM fresh, with updated version of MySQL, PHP, Apache, etc. I ask because I am not really the sysadmin, but have but put in charge of this server.

    Read the article

  • How can Windows 7 be set to sleep at 8pm and wake at 8am every day?

    - by Brett Alton
    I'm trying to figure out how to get Windows 7 to sleep at 8pm and wake at 8am on a daily basis. I'm looking at Windows' services and there is a Shutdown.exe that can run, but it only seems to have a /h flag for hibernation. For waking from sleep, there is a "wake Windows from sleep to run this task" option, but it never seems to work. There is also a Sleep.exe program I can download but it comes with 13MB of other libraries from Microsoft. Is there a C# program I can write to make this work?

    Read the article

  • where are user name and pasword saved in windows native vpn connection?

    - by Tîrêj
    I am using time freeze program but I can not save my free vpn user name and pasword. Since user name and pasword changed daily I have to unfreeze change pasword then freeze again. In the time freeze there is option to excusing file and folder. if I find the location of vpn setting that save the user name and pasword, I will have ability to use without unfreezing and re freezing. There is a location with path %userprofile%\AppData\Roaming\Microsoft\Network\Connections\Pbk\ that contain vpn configurations except user name and password. I need the location that store user name and password.

    Read the article

  • How to setup a scheduled task to run every 30 minutes? Windows Server 2003

    - by JL
    I'm having a problem running a scheduled task in Windows 2003 server, it ran for 4 hours, now it won't start up again. I will tell you exactly what I have done to create the task, please tell me if I am doing anything wrong. My objective is to have my task (console app) run every 30 minutes. I open control panel, scheduled tasks, add new scheduled task Browse for my application I want to run, a console app without any command line arguments I select, perform this task - Daily I select the start time (now 8:00 in the morning) Start date set to today Set the correct username + password Open advanced properties Repeat the task every 30 minutes for 24 hours Save and close. Anything wrong in the above mentioned steps I've implemented? Thanks in advance.

    Read the article

  • Why Does My Laptop Freeze Up When Docked?

    - by Michael Haren
    I have a Dell Latitude D520. First with Windows XP and now with Windows 7, it completely freezes/locks up if I dock it while it's awake. If I hibernate/sleep, dock, then wake/resume, everything's fine. I installed the Dell Notebook System Software for Vista (No Win7 version available) and updated the bios to the latest version but the problem persists. As long as I remember to put this thing in standby before I dock it, it's extremely reliable--it survives my daily commute and heavy use and typically only reboots a few times a month. Any other ideas for things I can try?

    Read the article

  • Hyper-v back-up

    - by Ddave23
    We are trying to decide a good backup strategy for our new Hyper-V setup. We have 3 VMs on Windows 2008 R2 Hyper-V host. We installed Symantec BackupExec 2010 on the host and have the Hyper-V Agent installed. We would like to perform a full backup at night to tape, and an incremental twice a day to a daily tape. Our environment needs constant protection for our database (Microsoft Access). Any thoughts? Should I be looking at different software?

    Read the article

  • Difference between "bit-number" in filesystem and that in OS (like 32,64 bit)?

    - by learner
    I encountered 2 terms ,"FAT32", a file system and "Windows Vista 32 bit". I found that the meaning of a 32 bit OS is that that the OS deals with data in chunks of minimum size 32 bits. I don't quite understand the depth of that, but I figure ,every file in that system with that OS should have a minimum size of 32 bits. I also read that these 32 bits are used to hold data of files' location(reference) and details. Which of it is it? I have also read that 4 GB of RAM is all that is needed at the most if you're on a 32 bit OS. But I don't understand why. If there are 32 bits to hold info about files and their locations,there can be 2^32 possible combinations of it. But I have found in many places,2^32 is divided by 1024 thrice to get 4GB. Why? Did that 2^32 become equal to 2^32 bytes? And about filesystems I read a similar explanation for what 32 means in FAT32. It is supposed to mean that 32 bits are used to number file system block. Now how is different from the number before the OS?

    Read the article

  • Can a mapped network drive be reconnected from the command line?

    - by Stephen Jennings
    On a daily basis I find myself in the Windows command prompt needing to access a network drive that is mapped but disconnected. I have yet to find a command that will reconnect this drive without unmapping and remapping (which leads to a password guessing game, since I don't own these computers). I would also like to be able to script this so every night the drive is reconnected if it has become disconnected somehow. The fastest solution I currently have is to: Type "start." to open explorer, Alt-D to focus the address bar, type the drive letter I want and press enter, and wait for it to display the drive contents, then finally, close explorer and go back to the command prompt. I know it's a minor inconvenience, but I'm often doing this through a slow VNC or PCAnywhere connection where doing anything through GUI is awful, so I'm just wondering if there's a better solution.

    Read the article

  • Large volume at /mnt on AWS instance

    - by rhaag71
    I know this is probably a somewhat 'dumb' question :) I have an AWS (small) instance and I just noticed that there is a ~150gb volume attached at /mnt, is this normal? It kinda freaked me out, I was thinking maybe someone was trying to capture whatever I mount in /mnt, there is the entry in my fstab too (and I found that others have this by googling)... the entry is as follows /dev/xvdb /mnt auto defaults,nobootwait,comment=cloudconfig 0 2 I don't have any volumes this large in my AWS volumes section though. I was just trying to understand this and be sure that someone is not trying to 'get in'... as there are many attempts daily. Thanks

    Read the article

  • Can anyone recommend free network monitoring tools?

    - by Josamoto
    Hi all I have a machine at home on a 3G internet connection, and my PC is consuming approximately about 200MB per day in uploads & downloads. I'm using an application called Networx to monitor the upload/download usage, but I am on the outlook for the culprit application munching away on my cap. Networx tells me how much my connection uses in total, and I need to know how much each application uploaded / downloaded. Does anyone know of a network connection monitoring utility for Windows 7 that can give me a daily outline of how much data was uploaded & downloaded, and specify this on a PER APPLICATION basis. Thanks in advance!

    Read the article

  • Can anyone recommend free network monitoring tools?

    - by Josamoto
    Hi all I have a machine at home on a 3G internet connection, and my PC is consuming approximately about 200MB per day in uploads & downloads. I'm using an application called Networx to monitor the upload/download usage, but I am on the outlook for the culprit application munching away on my cap. Networx tells me how much my connection uses in total, and I need to know how much each application uploaded / downloaded. Does anyone know of a network connection monitoring utility for Windows 7 that can give me a daily outline of how much data was uploaded & downloaded, and specify this on a PER APPLICATION basis. Thanks in advance!

    Read the article

  • How much visitors could handle my server ?

    - by coolboycsaba
    I have a website, and I want to host it on my own computer, but I'm wondering if it's good enough. The website check`s if the user is logged in and then displays 15 items (title, description) from a mysql database and the rating (stored in another database) and the comments (another database) for each item. It also displays some stats (number of items, comments). I also have an image for each item. My specs are: AMD Athlon 64 X2 Dual Core Processor 5600+ 2.90 GHz RAM memory: 4.00GB Windows 7 64bit So what do you think, how much visitors and items could it handle (at once or daily) ? My internet connection is good, around 7-10 mb upload and same download speed

    Read the article

< Previous Page | 100 101 102 103 104 105 106 107 108 109 110 111  | Next Page >