Search Results

Search found 13804 results on 553 pages for 'amazon elastic ip'.

Page 130/553 | < Previous Page | 126 127 128 129 130 131 132 133 134 135 136 137  | Next Page >

  • With a node.js powered server on EC2, how can I decrease the TCP connection time?

    - by talentedmrjones
    While profiling my application I've noticed that in the Firebug Net panel, the "Connecting" time—that is the time waiting for a TCP connection—is consistently around 70–100ms. See image below: Of course in the grand scheme of things, 100ms is not long, but I have seen other services that respond with 0ms Connect time. So if other servers can, I should be able to as well. Any thoughts on how I might even beging to troubleshoot this?

    Read the article

  • ec2-user password for running sudo -H -u

    - by bool.dev
    I have to run this command to initialize gitosis: sudo -H -u git gitosis-init < /home/ec2-user/id_rsa.pub But that asks me for a password for ec2-user: $ sudo -H -u git gitosis-init < id_rsa.pub [sudo] password for ec2-user: I do not have a password as i use the default .pem key file to login. I know i can probably login as the git user and do this, but is there any other way? Update: Using Linux AMI 12.09 (micro-instance), in region us-east-1 (N. Virginia)

    Read the article

  • How to find the reason for a weekly downtime on an Ubuntu web server hosted by AWS?

    - by IceSheep
    We started monitoring our web server using Pingdom and found out that we have a downtime of a few minutes every Sunday at 0:00 UTC. The test runs every minute and checks if a successful HTTP response (code 200) is returned on port 80. The test fails due to a timeout (no response after 30 seconds). Here's what we've already checked – without success: Since we run our webserver behind a load balancer, I've set the Pingdom test on the load balancer's public DNS and the webserver's public DNS in order to find out if there's a problem with the AWS load balancer – both tests return the same result We set up Munin on our webserver. Everything looked fine even after the failure. Since the last failure lasted only 2 minutes I suppose Munin couldn't capture a potential problem (it only checks every 5 minutes) I have checked /var/log/apache2/error.log and /var/log/syslog for suspicious entries I have checked /etc/cron.weekly and /etc/crontab for suspicious entries I have searched for files created or last-modified during 0:00 and 0:15 using this method: touch -t 201209020000 start touch -t 201209020015 end find / -newer start -and ! -newer end (nothing found) Has anybody experienced a similar problem? Any proposals on how to find the reason for this behavior? It's Ubuntu 10.04 LTS running on an AWS m1.large instance. Thanks!

    Read the article

  • How to keep multiple servers in sync file wise?

    - by GForceSys
    I'm currently managing a cluster of PHP-FPM servers, all of which tend to get out of sync with each other. The application that I'm using on top of the app servers (Magento) allows for admins to modify various files on the system, but now that the site is in a clustered set up modifying a file only modifies it on a single instance (on one of the app servers) of the various machines in the cluster. Is there an open-source application for Linux that may allow me to keep all of these servers in sync? I have no problem with creating a small VM instance that can listen for changes from machines to sync. In theory, the perfect application would have small clients that run on each machine to be synced, which would talk to the master server which would then decide how/what to sync from each machine. I have already examined the possibilities of running a centralized file server, but unfortunately my app servers are spread out between EC2 and physical machines, which makes this unfeasible. As there are multiple app servers (some of which are dynamically created depending on the load of the site), simply setting up a rsync cron job is not efficient as the cron job would have to be modified on each machine to send files to every other machine in the cluster, and that would just be a whole bunch of unnecessary data transfers/ssh connections.

    Read the article

  • Need to scale quickly. Which cloud service should I use?

    - by mk1000
    Traffic to my facebook app is growing at an insane rate and I need some suggestions on how to scale. I'm probably not going to even be able to keep it running by the day's end, as it's hosted from my already overloaded dedicated server. I need to either move it to its own box or a cloud service like e2c. Something like e2c seems like the way to go, but my server admin skills are terrible. Is there a good front end management UI for e2c or another hosting service that is comparable in cost that is fully managed? I don't mind going with something a bit more expensive now if that means I can get everything switched over and running within 24 hours.

    Read the article

  • options for physical architecture of rails site regarding caching server or cdn

    - by timpone
    I have a rails app that is set on a single server currently. On production, I force_ssl for everything. I am interested in using a caching server for images (I'm fine with css and js being served from origin for time being). Would nginx or varnish (which I have no experience with) be a better solution (for October 2012)? I'd imagine that it would be easy to switch these around while still on this single server architecture. Or would something like cloudfront (which I also have no experience with) make sense for hosting image files? I know this is a vague question but appreciate any current feedback. thx in advance

    Read the article

  • how to fight back attacks on my web service

    - by user12145
    My apache webservice is getting a large quantity of requests over the days, each one with somewhat random login to gain access. I identified about 60 such ips(a few sample below), all belong to google. is there a way to find more information about the origin of the attacker? or should I just block these ips. secondly, should I attempt to block the identified ips subnets(74.125.46.*) as a preventive measure? 72.14.194.65 64.233.172.20 74.125.75.19 72.14.194.33 74.125.46.87 74.125.44.91 74.125.46.91

    Read the article

  • Large volume at /mnt on AWS instance

    - by rhaag71
    I know this is probably a somewhat 'dumb' question :) I have an AWS (small) instance and I just noticed that there is a ~150gb volume attached at /mnt, is this normal? It kinda freaked me out, I was thinking maybe someone was trying to capture whatever I mount in /mnt, there is the entry in my fstab too (and I found that others have this by googling)... the entry is as follows /dev/xvdb /mnt auto defaults,nobootwait,comment=cloudconfig 0 2 I don't have any volumes this large in my AWS volumes section though. I was just trying to understand this and be sure that someone is not trying to 'get in'... as there are many attempts daily. Thanks

    Read the article

  • Passive mode FTP file download hangs from specific machine

    - by chiptuned
    I have a server which is an AWS instance that just cannot download files from a specific FTP server. I can connect to the FTP server fine and run some commands, but when I request a file it just hangs. Here is the debug output of the base linux ftp client after login: ---> SYST 215 UNIX Type: Apache FtpServer Remote system type is UNIX. ftp> get outgoing/catalog.gz catalog.gz local: catalog.gz remote: outgoing/catalog.gz ---> PASV 227 Entering Passive Mode (64,156,167,125,135,191) ---> RETR outgoing/catalog.gz 150 File status okay; about to open data connection. Thats it. Then it just sits there and nothing transfers. I have verified that a data connection is made but the client gets no data. ? ss -nt dst 64.156.167.125 State Recv-Q Send-Q Local Address:Port Peer Address:Port ESTAB 0 0 10.185.147.150:41190 64.156.167.125:21 ESTAB 0 0 10.185.147.150:48871 64.156.167.125:48557 The FTP server is not in my control and downloads from other FTP servers in passive mode have worked. Active mode does not work as the system is behind a firewall. Every FTP client I've tried has the same problem. The download works from other systems, even from other AWS instances I have with the same Security Group. Not necessarily the same distro or config though. I understand it may be some issue on the server side, but I want to know what it is about my particular machine where the transfer hangs and where on every other machine I can get my hands on, it works. Please let me know what the culprit on the client side could be or ideas on what else to look at.

    Read the article

  • Why would one server be sending DUP ACK packets to one PC, which is responding with HTTP RST packets?

    - by IronicMuffin
    I'm not a network profressional, so please excuse any wrong language. I was debugging why my DNS traffic was a constant 160Kbps on our corporate network. I opened up a wireshark trace, and I see one PC of a coworker broadcasting HTTP [RST] packets to one of our DMZ servers at the rate of 1000 a second. He restarted his machine, and as soon as it went offline, the server started broadcasting [DUP] [ACK] packets, until he came back online. It then resumed the HTTP [RST] packets. Apparently this server has been doing this kind of behavior since it went live. I believe it did this with a printer and an access point as well. Can anyone explain why this behavior is occurring? Any solutions? The initial research was done because there have been "bandwidth issues" and I wonder if this is contributing.

    Read the article

  • Finding the owner of an AWS access key + secret key pair

    - by nightw
    I would like to have a simple solution (possibly in 1-3 plain API calls to AWS) to find the owner of an AWS access key. I have the password of the "root" AWS account and of course I can manage the users and credentials through IAM, but we have a lot of users and I don't want to look at them one by one looking for the owner of the key. So basically I have a working access key + secret key pair (in fact a couple of them), but I do not know which user's key is it and what rights are on it. What is the easiest way to do this? Thank you in advance.

    Read the article

  • How to configure S3 or DNS to handle incomplete name (sans www) for web site?

    - by user193116
    I have a set up a bucket called "www.mydomainname.com" to host my website and I have configured the CNAME such that "www.mydomainname.com" points to the my endopint http://www.mydomainname.com.s3-website-us-east-1.amazonaws.com/ It works and when people who type the the full url "www.mydomainname.com" are able to see my index page But most people are in the habit of typing incoplete domain name -- they just type "mydomainname.com" and their browser fails to find my site. Is there a way to configure CName or S3 bucket such that typing "mydomainname.com" take them to my s3 website ? (I am using Networksolutions as my DNS provider).

    Read the article

  • FTP Server with MySQL access, and POST notification

    - by TIW
    Im looking for an FTP server solution, that we can host either internally on a dedicated server, or on Rackspace Cloud/AWS, that provides a HTTP POST notification when a file is uploaded, and allows user accounts to be created either through an API or MySQL database. There are several offerings that provide email notification - but has anyone come across anything that matches the above requirements. BrickFTP being a IaaS system is an option, but we would prefer something hosted in house. I don't believe the standard FTP servers provided with Apache can do the above ... can they?

    Read the article

  • How to route traffic via another machine before the default gateway

    - by Rich
    At the moment I have a router on 192.168.0.1, a Linux box on 192.168.0.2 and desktop clients from 192.168.0.3. Everything works with 192.168.0.1 as the default gateway. I'd like to send the traffic from the desktop clients via the Linux box before it goes out through the router so I can sniff the traffic (some of these are wireless connections). Can I set the default gateway to 192.168.0.2 on the desktop clients and then perhaps add some iptables rules to forward this traffic through 192.168.0.1? Quite happy to change the client desktops to another subnet if that makes it easier. Thanks in advance.

    Read the article

  • What's the best practice for taking MySQL dump, encrypting it and then pushing to s3?

    - by HalogenCreative
    This current project requires that the DB be dumped, encrypted and pushed to s3. I'm wondering what might be some "best practices" for such a task. As of now I'm using a pretty straight ahead method but would like to have some better ideas where security is concerned. Here is the start of my script: mysqldump -u root --password="lepass" --all-databases --single-transaction > db.backup.sql tar -c db.backup.sql | openssl des3 -salt --passphrase foopass > db.backup.tarfile s3put backup/db.backup.tarfile db.backup.tarfile # Let's pull it down again and untar it for kicks s3get surgeryflow-backup/db/db.backup.tarfile db.backup.tarfile cat db.backup.tarfile | openssl des3 -d -salt --passphrase foopass |tar -xvj Obviously the problem is that this script everything an attacker would need to raise hell. Any thoughts, critiques and suggestions for this task will be appreciated.

    Read the article

  • mirror sql server 2008 to AWS instance from our datacenter?

    - by Alex
    We are currenlty running on hosted pos system locally and would like to mirror to AWS. We are new to AWS and would like to know the most cost effective way to do this? We have 2 DB and 2 web servers right now in one cabinet in CA. One tape drive, one firewall, one SNA. We are thinking to replicate our system in AWS (using sql server 2008) and just mirror both systems and use a witness server between them to keep the data in sync? The goal is, if CA datacenter goes down, AWS keeps running. User see no downtime. All data is synced. Is anyone doing something similar? Would this be practical to just use AWS in this fashion? Thanks

    Read the article

  • 1 Computer Pulls Wrong DNS Every Morning

    - by bradenkeith
    One of my users takes his laptop home, connects to his internal network fine. He brings it in every morning and since Monday - every morning it pulls the wrong DNS every time. It pulls 1 DNS: 192.168.1.1 which is the router. The DNS are external IPs. I have several users that take their laptops home and connect to other networks - what could cause this anomaly? He's booting up when he comes in the office and a simple ipconfig /release /renew fixes the issue.

    Read the article

  • How to move a Windows instance to a different EC2 data centre?

    - by Darren Cook
    I've a Windows EC2 instance running in one region and need to move it to another region (Tokyo and Singapore in this case). Is that even possible? What potential problems do I need to watch out for? (I found http://stackoverflow.com/questions/2181849/ec2-instance-cloning which is describing how to do it, but appears to assume Linux instances, and to assume the same data centre. Is it possible to move my keys across to another region?) I tried something similar with a Windows instance a few months ago, just trying to clone it in the same data centre, but I couldn't get it working quickly, so I had to give up and just create a fresh instance at that time. This time I've got a bit of breathing space, and want to research how to do it properly! Root Device Type: ebs Block Devices: sda1 xvdf (both are ebs, "attached", and have Delete on termination set to "no"; sda1 is the root device) The AMI is described as "Unavailable" (then an ami number).

    Read the article

  • Is there a way to send personal documents on Kindle for Mac app?

    - by Sid
    I have the Kindle App on my Mac, and an android phone. When I email documents to my [email protected] id, I am able to see it in my library, and subsequently send it to my Android device. However, I'm not able to send it to the Kindle App for my Mac. The Kindle for Mac FAQs clearly state that Magazines, personal documents, etc. are not supported. However, I came across here that there is a workaround to this, although I've not been able to figure out what it is.

    Read the article

  • What's the best approach when it comes to updating a production(on ec2) machine that can't go down?

    - by Ryan Detzel
    We have three main servers on ec2, web, database, and search. I logged in today to find: 77 packages can be updated. 45 updates are security updates. which scares the crap out of me so I want to update these machines asap but I'm scared to just run the updates on a live running system. Is this safe to do, what's the best approach when it comes to doing security updates on production machines?

    Read the article

  • Recommend a desktop app to monitor port 80 on servers

    - by busyone
    Hello, I did some googling but was unsuccessful, so I am posting here. I am looking to get an app (preferably free/open source) to monitor my servers on port 80 that would send me an email/text when some predefined timeout triggers. I could probably write something in VB to do this but i am burried with projects, why reinvent the wheel. I know there are services out there that want you to pay $5/month to do so, I was thinking about a Windows app that would sit on my computer and simply monitor predefined list of IPs on port 80. Thanks!

    Read the article

  • Easiest way to get feedback from an EC2 instance

    - by Sanity
    I need to run a script on an ec2 instance once a day, and I'd like to have some easy way for it to let me know if something went wrong. I would prefer not to have to modify the original image, which is a recent version of Ubuntu, so ideally I'd like to do all setup in the script I pass to the Ubuntu instance through the ec2-run-instances command. I've considered creating a gmail account for it, and letting it send email through that - but the setup was rather involved, with certificates and such things. I've looked at using the gist API, but anything uploaded through it is public. The Google command line tool also appears quite complicated to set up. Is there some easier way to do this?

    Read the article

  • My visitors %20 percent can't reachable to my web site

    - by user39110
    Hi, i have dedicated web server and also i hosting www.btgmaslak.com web site. I'm facing a problem about 1 days. My visitors about 20 percent can't reach my web site. I thing this problem occurs because of dns settings. Can you analyze my dns settings and give me suggestions ? thanks http://www.intodns.com/btgmaslak.com

    Read the article

< Previous Page | 126 127 128 129 130 131 132 133 134 135 136 137  | Next Page >