Search Results

Search found 37180 results on 1488 pages for 'proxy pass request failed'.

Page 20/1488 | < Previous Page | 16 17 18 19 20 21 22 23 24 25 26 27  | Next Page >

  • Reverse proxy with SSL and IP passthrough?

    - by Paul
    Turns out that the IP of a much-needed new website is blocked from inside our organization's network for reasons that will take weeks to fix. In the meantime, could we set up a reverse proxy on an Internet-based server which will forward SSL traffic and perhaps client IPs to the external site? Load will be light. No need to terminate SSL on the proxy. We may be able to poison DNS so original URL can work. How do I learn if I need URL rewriting? Squid/apache/nginx/something else? Setup would be fastest on Win 2000, but other OSes are OK if that would help. Simple and quick are good since it's a temporary solution. Thanks for your thoughts!

    Read the article

  • Proxy server on windows with SSL encrypted exchange with client

    - by Syffys
    I want to set up a classic proxy server (HTTPS, HTTPS, SSH, FTP, etc...) on a windows platform, but I need the following features: password authentication for clients data exchanged between clients and server to be SSL encrypted I've been trying to set this up various application to get this result, but without success so far: Squid for windows ccproxy wingate Alternatively, an other solution would be an HTTP SSL tunnel encapsulating an unencrypted proxy connection between clients and the server. I've spent a lot of time without any result so far, so I'm wondering if anyone faced this kind of issue. Thanks in advance!

    Read the article

  • Allowing Skype through Squid proxy

    - by Blue Gene
    I have a squid proxy running on 192.168.1.2 on port 3000. with Squid i am able to connect to internet and browse websites but skype is not connecting even after i specified proxy server in it. In the skype configuration i mentioned it to use SOCKS5 through host 192.168.1.2 and port 3000. But still its not working.Port 443 is open in squid configuration. https is working ok as i can access gmail and bank payment sites,but skype can not access its server. acl Safe_ports port 443 # https acl SSL_ports port 443 also tried setting acl numeric_IPs dstdom_regex ^(([0-9]+\.[0-9]+\.[0-9]+\.[0-9]+)|(\[([0-9af]+)?:([0-9af:]+)?:([0-9af]+)?\])):443 acl Skype_UA browser ^skype http_access allow CONNECT numeric_IPS Skype_UA IPTABLES is set to accept all

    Read the article

  • Reverse Proxy Server SSL?

    - by valveLondon
    Context We currently have an Apache web server in the DMZ set up as a reverse proxy and load balancer for two machines running Windows Server 2008 (IIS) inside. The Apache server has a genuine SSL certificate and serves up both http and https, however, the balancer members in the load balancing section are set to: BalancerMember {https://server1} and {https://server2}. The IIS web servers have self-signed certificates in order to respond to the https requests. My question: Do we need to forward any requests from Apache (in the DMZ) to the inside using SSL? e.g can the reverse proxy forward the requests using HTTP? and if so, why would I choose to forward them with SSL? (how secure is the http line between the dmz and the inside); In other words, can I totally disable SSL on my inside web servers?

    Read the article

  • Apache reverse proxy setup

    - by nixnotwin
    I have a jboss application server on machine1. The application address is http://ip-address:8080/webapp. I wanted to have only an ip pointing to the application. So on machine2 I setup an apache proxy. But it only helps to shift to port 80 but the directory webapp cannot be removed. So using proxy, the address is http://ip-address/webapp. So is there a way to just have the ip point to the application. For example the address http://ip-address should open the web page of the application.

    Read the article

  • Auto-detect proxy settings for the network

    - by user42891
    Firefox browser contains network settings under Tools--Options--Advanced--Network--Settings and there is an option to do auto detect proxy settings, how should I enable this? Currently this is manually configured and its possible for users to bypass and use the internet directly. We use a variety of browsers (firefox, IE, chrome, safari, opera) on win xp, win 2003, win vista machines. How should I enable this so that the end user cannot manipulate the settings on his browser to by pass security. I have configured a squid cache proxy server for this purpose.

    Read the article

  • Run SSH trough a proxy

    - by Row Minds
    I have a question. How can i run SSH trough a proxy or a SSH Tunneling Protocol? For example i have a computer(Ubuntu) with ip eg. 123.123.123.123 and i want to connect to my server trough a proxy eg. 111.111.111.111 so i can see in lastlog that the last authentication was made from 123.123.123.123 . I need this because i cannot access a certain server at work only with an specified range IP. What i tried so far was http://daniel.haxx.se/docs/sshproxy.html where i used ssh -D 1 [email protected] -p 443 where i had 127.0.0.1 forwarded to a SSH Tunneling Protocol http://www.nixtutor.com/linux/installing-and-configuring-an-ssh-server/ (i configurated listening on port 443 but no result, still no connection (denied) ...) This question may sound simple, but i can't figure it out, can you please help me? Thanks.

    Read the article

  • Setting up a reverse proxy [on hold]

    - by mrwooster
    I am looking for the best solution for setting up a very low maintenance reverse proxy for a production website (example.com). The setup is as follows: A blog with will be hosted on heroku, and will reside at example.com/blog A static info page which will be hosted on S3 and will reside at example.com/signup A dynamic content management system, provided and hosted by an external vendor which will respond to requests for any other pages. The two solutions which come to mind are: Use HAProxy Ask the external vendor to reverse proxy requests for /blog and /signup The obvious solution would be to use HAProxy, but, if at all possible, I would like to avoid having to setup and maintain another server (especially such a critical one). I came across a company called Snapt which offers hosted HAProxy solutions, but its more geared towards load-balancing than reverse proxying. Option 2 is a possibility, but gives us very little control over changes and configuration. I see a lot of sites hosting blogs on /blog so this must be fairly common practice, am I missing an obvious solution?

    Read the article

  • iphone - Programmatically set (System-wide) proxy settings?

    - by Andrew
    I am new to iPhone development, so I'm sorry if this is a stupid question. I am developing an application whose purpose will be to route all iPhone activity through my company's proxy. Is there a way to programmatically set system-wide proxy settings in the iPhone (which will also take effect on the 3G connection)? I know there is a way to manually set proxy settings for each wifi connection. Detecting new networks and setting the proxy on them would be acceptable. However, I need to also be able to set the proxy on the 3G connection. Also, bonus: Is there a way to programmatically change the "Restrictions" settings? If anyone has any tips or can point me in the right direction, I would appreciate it. Thanks. EDIT: Please understand that this is for a legitimate purpose. Apple has to approve app store additions, so it's not like I'm trying to spread a virus. Please, constructive answers only.

    Read the article

  • slsvcutil.exe Proxy and Interfaces

    - by JPM
    Is it possible when using slsvcutil.exe to generate a proxy through the command line not to have the proxy file output the Interface in an Asynchronous fashion. For example, if I have a function "foo()" on the serverside in the Interface, when I generate the proxy using Slsvcutil.exe, it makes two functions in the interface definition in the proxy named "BeginFoo()" and "EndFoo()". All I want is "Foo()", I don't need the other two methods. Is this possible? I'm using the proxy with Monotouch which is why I need to use Slsvcutil.exe but don't need the Asynchronous methods. Thanks!

    Read the article

  • App Engine webapp.RequestHandler child instances has no self.request during __init__

    - by grucha
    i use modified webapp.RequestHandler for handling requests in my app: class MyRequestHandler(webapp.RequestHandler): """ Request handler with some facilities like user. self.out is the dictionary to pass to templates """ def __init__(self, *args, **kwargs): super(MyRequestHandler, self).__init__(*args, **kwargs) self.out = { 'user': users.get_current_user(), 'logout_url': users.create_logout_url(self.request.uri) } def render(self, template_name): """ Shortcut to render templates """ self.response.out.write(template.render(template_name, self.out)) class DeviceList(MyRequestHandler): def get(self): self.out['devices'] = GPSDevice.all().fetch(1000) self.render('templates/device_list.html') but I get an exception: line 28, in __init__ self.out['logout_url'] = users.create_logout_url(self.request.uri) AttributeError: 'DeviceList' object has no attribute 'request' When the code causing exception is moved out of __init__ everything's fine: class MyRequestHandler(webapp.RequestHandler): """ Request handler with some facilities like user. self.out is the dictionary to pass to templates and initially it contains user object for example """ def __init__(self, *args, **kwargs): super(MyRequestHandler, self).__init__(*args, **kwargs) self.out = { 'user': users.get_current_user(), } def render(self, template_name): """ Shortcut to render templates """ self.out['logout_url'] = users.create_logout_url(self.request.uri) self.response.out.write(template.render(template_name, self.out)) Whi is that? Why there's no self.request after parent's (i.e. webapp.RequestHandler's) __init__ was executed?

    Read the article

  • Consume restful webservice through web proxy

    - by Chico
    Hello, I'm trying to consume a restful webservice in java using the Apache Wink framework through my school web proxy requiring authentification ClientConfig clientConfig = new ClientConfig(); clientConfig.proxyHost("proxy.school.com"); clientConfig.proxyPort(3128); //nothing to set username and password :( RestClient client = new RestClient(clientConfig); Resource resource = client.resource("http://vimeo.com/api/v2/artist/videos.xml"); String response = resource.accept("text/plain").get(String.class); I've also tried to use the BasicAuthSecurityHandler but it seems to be used to authenticate directly to a web server, not the web proxy BasicAuthSecurityHandler basicAuthHandler = new BasicAuthSecurityHandler(); basicAuthHandler.setUserName("username"); basicAuthHandler.setPassword("password"); config.handlers(basicAuthHandler); It still fail with a HTTP 407 error code : Proxy Authentication Required. I've googled the best I could, nothing came up better to consume a webservice from a Java client through a web proxy, if someone has another idea, feel free to respond

    Read the article

  • Huge amount of time sending data with suds and proxy

    - by Roman
    Hi everyone, I have the following code to send data through a proxy using suds: import suds t = suds.transport.http.HttpTransport() proxy = urllib2.ProxyHandler({'http': 'http://192.168.3.217:3128'}) opener = urllib2.build_opener(proxy) t.urlopener = opener ws = suds.client.Client('http://xxxxxxx/web.asmx?WSDL', transport=t) req = ws.factory.create('ActionRequest.request') req.SerialNumber = 'asdf' req.HostName = 'hola' res = ws.service.ActionRequest(req) I don't know why, but it can be sending data above 2 or 3 minutes, or even more and it raises a "Gateway timeout" exception sometimes. If I don't use the proxy, the amount of time used is above 2 seconds or less. Here is the SOAP reply: (ActionResponse){ Id = None Action = "Action.None" Objects = "" } The proxy is running right with other requests through urllib2, or using normal web browsers like firefox. Does anyone have any idea what's happening here with suds? Thanks a lot in advance!!!

    Read the article

  • How to make proxy on nginx?

    - by Adrian K.
    How would I set my webservers to work in way described below? Http request: mypublic.com --- handled normally by nginx as it is set up already (listen 80;) Http request: myprivate.com --- handled by apache set up to work on 8080 (listen 8080) I'd like to avoid including ports when typing address in browser, some kind of mockup (proxy?). Both of domains are pointing to my machine and set up by named.

    Read the article

  • Nginx Reverse Proxy : post_action if proxy cache hit - Possbile?

    - by anonymous-one
    We have recently found out about nginxes post_action. We were wondering it there was a way to use this directive if a proxy cache hit is made? The flow we were hoping on is as follows: 1) User request comes in 2) If cache HIT goto A / If cache MISS goto B A) 1) Serve Cached Result A) 2) post_action to another url on the backend B) 1) Server request from backend B) 2) Store result from backend Any ideas if this is possible via post_action? Thanks!

    Read the article

  • How to invoke a command using specific proxy server?

    - by Xiè Jìléi
    Some applications support proxy (http proxy or socks proxy), and some are not. For browsers, I can specify proxy server in the preferences/options dialog, and other applications may be able to configure proxy servers in config files. For general purpose, can I invoke a command using a specific proxy? Like following: $ proxy-exec --type sock5 --server 1.2.3.4:8000 -- wget/ftp ... I'm using Ubuntu Maverick. P.S. In win32, it can be implemented by hijacking the socket dlls, maybe, I'm not familiar with Linux programming, but I guess it's possible in Linux. though.

    Read the article

  • Multiple connections in a single SSH SOCKS 5 Proxy

    - by Elie Zedeck
    Hey guys, My fist question here on Stackoverflow: What should I need to do so that the SSH SOCKS 5 Proxy (SSH2) will allow multiple connections? What I have noticed, is that when I load a page in Firefox (already configured to use the SOCKS 5 proxy), it loads everything one by one. It can be perceived by bare eyes, and I also confirm that through the use of Firebug's NET tab, which logs the connections that have been made. I have already configure some of the directives in the about:config page, like pipeline, persistent proxy connections, and a few other things. But I still get this kind of sequential load of resources, which is noticeably very slow. network.http.pipelining;true network.http.pipelining.maxrequests;8 network.http.pipelining.ssl;true network.http.proxy.pipelining;true network.http.max-persistent-connections-per-proxy;100 network.proxy.socks_remote_dns;true My ISP sucks because during the day, it intentionally breaks connections on a random basis. And so, it is impossible to actually accomplish meaningful works without the need of a lot of browser refresh or hitting F5 key. So, that is why I started to find solutions to this. The SSH's dynamic port forwarding is the best solution I find to date, because it has some pretty good compression which saves a lot of useless traffic, and is also secure. The only thing remaining is to get it to have multiple connections running in it. Thanks for all the inputs.

    Read the article

  • How to Send Headers Through PHP Proxy?

    - by blabus
    Okay, so I have an issue with an AJAX request. I currently have this URL: http://www.google.com/images?hl=en&safe=off&gbv=2&tbs=isch%3A1&sa=1&q=cars+imagesize%3A500x500&aq=f&aqi=&aql=&oq=&gs_rfai=&start=0 I then pass it to my proxy script by modifying the string to this: proxy.php?url=http://www.google.com/images?hl=en&safe=off&gbv=2&tbs=isch%3A1&sa=1&q=cars+imagesize%3A500x500&aq=f&aqi=&aql=&oq=&gs_rfai=&start=0 I need to use the PHP proxy script to grab that page's HTML so that I can then parse through it with javascript. However, the problem is that the headers in that URL are also being sent to the proxy script, and as a result, I get a 'malformed or illegal request' error. I'm pretty sure the two different sets of headers are the problem, because if I just replace the original URL string with 'http://wwww.google.com', the proxy returns the HTML of the page correctly. So basically, I don't know how to fix this. I'm a complete PHP noob, and I tried escaping the original URL before I appended it to the 'proxy.php?url=', but that doesn't fix anything. Any ideas? Thanks!

    Read the article

  • Can't upgrade ubuntu 9.xx to 12.04

    - by andrej spyk
    I can't upgrade old Ubuntu 9.10 to new, if I check for upgrade it says: Could not download all repository indexes *Failed to fetch ttp://security.ubuntu.com/ubuntu/dists/jaunty-security/main/binary-i386/Packages 404 Not Found Failed to fetch ttp://security.ubuntu.com/ubuntu/dists/jaunty-security/restricted/binary-i386/Packages 404 Not Found Failed to fetch ttp://security.ubuntu.com/ubuntu/dists/jaunty-security/main/source/Sources 404 Not Found Failed to fetch ttp://security.ubuntu.com/ubuntu/dists/jaunty-security/restricted/source/Sources 404 Not Found Failed to fetch ttp://cz.archive.ubuntu.com/ubuntu/dists/jaunty/main/binary-i386/Packages 404 Not Found Failed to fetch ttp://cz.archive.ubuntu.com/ubuntu/dists/jaunty/restricted/binary-i386/Packages 404 Not Found Failed to fetch ttp://cz.archive.ubuntu.com/ubuntu/dists/jaunty/main/source/Sources 404 Not Found Failed to fetch ttp://security.ubuntu.com/ubuntu/dists/jaunty-security/universe/binary-i386/Packages 404 Not Found Failed to fetch ttp://security.ubuntu.com/ubuntu/dists/jaunty-security/universe/source/Sources 404 Not Found Failed to fetch ttp://cz.archive.ubuntu.com/ubuntu/dists/jaunty/restricted/source/Sources 404 Not Found Failed to fetch ttp://cz.archive.ubuntu.com/ubuntu/dists/jaunty/universe/binary-i386/Packages 404 Not Found Failed to fetch http://cz.archive.ubuntu.com/ubuntu/dists/jaunty/universe/source/Sources 404 Not Found Failed to fetch ttp://security.ubuntu.com/ubuntu/dists/jaunty-security/multiverse/binary-i386/Packages 404 Not Found Failed to fetch ttp://cz.archive.ubuntu.com/ubuntu/dists/jaunty/multiverse/binary-i386/Packages 404 Not Found Failed to fetch tp://cz.archive.ubuntu.com/ubuntu/dists/jaunty/multiverse/source/Sources 404 Not Found Failed to fetch htp://cz.archive.ubuntu.com/ubuntu/dists/jaunty-updates/main/binary-i386/Packages 404 Not Found Failed to fetch ttp://cz.archive.ubuntu.com/ubuntu/dists/jaunty-updates/restricted/binary-i386/Packages 404 Not Found Failed to fetch ttp://security.ubuntu.com/ubuntu/dists/jaunty-security/multiverse/source/Sources 404 Not Found Failed to fetch ttp://cz.archive.ubuntu.com/ubuntu/dists/jaunty-updates/main/source/Sources 404 Not Found Failed to fetch ttp://cz.archive.ubuntu.com/ubuntu/dists/jaunty-updates/restricted/source/Sources 404 Not Found Failed to fetch http://cz.archive.ubuntu.com/ubuntu/dists/jaunty-updates/universe/binary-i386/Packages 404 Not Found Failed to fetch ttp://cz.archive.ubuntu.com/ubuntu/dists/jaunty-updates/universe/source/Sources 404 Not Found Failed to fetch ttp://cz.archive.ubuntu.com/ubuntu/dists/jaunty-updates/multiverse/binary-i386/Packages 404 Not Found Failed to fetch ttp://cz.archive.ubuntu.com/ubuntu/dists/jaunty-updates/multiverse/source/Sources 404 Not Found Some index files failed to download, they have been ignored, or old ones used instead.* How I can upgrade if I can't burn new CD?

    Read the article

  • csync2 ERROR: Connection to remote host failed

    - by Emil Salama
    I was unable to find any articles to answer this question, so my best bet was to post this here: Scenario We have 2x application servers in production hosting a PHP website and I would like some folders to be syncronized between the 2, the same was setup for the development environment with no issues, I've followed all instructions from the URL "http://www.cloudedify.com/synchronising-files-in-cloud-with-csync2/", I still seem to have the same result, firewall has been disabled on both boxes for troubeshooting purposes: Config Files: cysnc2.cfg nossl * *; group production { host server1; host server2; key /etc/csync-production-group.key; include /etc/httpd/sites-available; include /xxxxxx/public_html/files include /xxxxxxx/magento/media/catalog/product include /xxxxxxx/magento/media/brands exclude *.log; exclude /xxxx/public_html/file/cache; exclude /xxxxx/public_html/magento/var/cache; exclude /xxxx/public_html/logs; exclude /xxxxx/public_html/magento/var/log; backup-directory /data/sync-conflicts/; backup-generations 2; auto younger; } /etc/xinetd.d/csync2 csync2.cfg service csync2 { disable = no flags = REUSE socket_type = stream wait = no user = root group = root server = /usr/sbin/csync2 server_args = -i -D /data/sync-db/ port = 30865 type = UNLISTED log_type = FILE /data/logs/csync2/csync2-xinetd.log log_on_failure += USERID } I've made sure that the daemon is listening on both server on port 30865 and the keys matched on both servers I've run a tcpdump on each server, output as follows: 12:20:31.366771 IP server1.49919 server2.csync2: Flags [S], seq 445156159, win 14600, options [mss 1460,sackOK,TS val 794864936 ecr 0,nop,wscale 7], length 0 12:20:31.366810 IP server2.csync2 server1.49919: Flags [S.], seq 450593575, ack 445156160, win 14480, options [mss 1460,sackOK,TS val 794798911 ecr 794864936,nop,wscale 7], length 0 12:20:31.367101 IP server1.49919 server2.csync2: Flags [.], ack 1, win 115, options [nop,nop,TS val 794864937 ecr 794798911], length 0 12:20:31.367138 IP server1.49919 server2.csync2: Flags [P.], seq 1:9, ack 1, win 115, options [nop,nop,TS val 794864937 ecr 794798911], length 8 12:20:31.367147 IP server2.csync2 server1.49919: Flags [.], ack 9, win 114, options [nop,nop,TS val 794798912 ecr 794864937], length 0 12:20:31.368625 IP server2.csync2 server1.49919: Flags [R.], seq 1, ack 9, win 114, options [nop,nop,TS val 794798913 ecr 794864937], length 0 Is there anything else i'm missing or should be doing?

    Read the article

  • Ubuntu 9.10 and Squid 2.7 Transparent Proxy TCP_DENIED

    - by user38400
    Hi, We've spent the last two days trying to get squid 2.7 to work with ubuntu 9.10. The computer running ubuntu has two network interfaces: eth0 and eth1 with dhcp running on eth1. Both interfaces have static ip's, eth0 is connected to the Internet and eth1 is connected to our LAN. We have followed literally dozens of different tutorials with no success. The tutorial here was the last one we did that actually got us some sort of results: http://www.basicconfig.com/linuxnetwork/setup_ubuntu_squid_proxy_server_beginner_guide. When we try to access a site like seriouswheels.com from the LAN we get the following message on the client machine: ERROR The requested URL could not be retrieved Invalid Request error was encountered while trying to process the request: GET / HTTP/1.1 Host: www.seriouswheels.com Connection: keep-alive User-Agent: Mozilla/5.0 (X11; U; Linux x86_64; en-US) AppleWebKit/532.9 (KHTML, like Gecko) Chrome/5.0.307.11 Safari/532.9 Cache-Control: max-age=0 Accept: application/xml,application/xhtml+xml,text/html;q=0.9,text/plain;q=0.8,image/png,/;q=0.5 Accept-Encoding: gzip,deflate,sdch Cookie: __utmz=88947353.1269218405.1.1.utmccn=(direct)|utmcsr=(direct)|utmcmd=(none); __qca=P0-1052556952-1269218405250; __utma=88947353.1027590811.1269218405.1269218405.1269218405.1; __qseg=Q_D Accept-Language: en-US,en;q=0.8 Accept-Charset: ISO-8859-1,utf-8;q=0.7,*;q=0.3 Some possible problems are: Missing or unknown request method. Missing URL. Missing HTTP Identifier (HTTP/1.0). Request is too large. Content-Length missing for POST or PUT requests. Illegal character in hostname; underscores are not allowed. Your cache administrator is webmaster. Below are all the configuration files: /etc/squid/squid.conf, /etc/network/if-up.d/00-firewall, /etc/network/interfaces, /var/log/squid/access.log. Something somewhere is wrong but we cannot figure out where. Our end goal for all of this is the superimpose content onto every page that a client requests on the LAN. We've been told that squid is the way to do this but at this point in the game we are just trying to get squid setup correctly as our proxy. Thanks in advance. squid.conf acl all src all acl manager proto cache_object acl localhost src 127.0.0.1/32 acl to_localhost dst 127.0.0.0/8 acl localnet src 192.168.0.0/24 acl SSL_ports port 443 # https acl SSL_ports port 563 # snews acl SSL_ports port 873 # rsync acl Safe_ports port 80 # http acl Safe_ports port 21 # ftp acl Safe_ports port 443 # https acl Safe_ports port 70 # gopher acl Safe_ports port 210 # wais acl Safe_ports port 1025-65535 # unregistered ports acl Safe_ports port 280 # http-mgmt acl Safe_ports port 488 # gss-http acl Safe_ports port 591 # filemaker acl Safe_ports port 777 # multiling http acl Safe_ports port 631 # cups acl Safe_ports port 873 # rsync acl Safe_ports port 901 # SWAT acl purge method PURGE acl CONNECT method CONNECT http_access allow manager localhost http_access deny manager http_access allow purge localhost http_access deny purge http_access deny !Safe_ports http_access deny CONNECT !SSL_ports http_access allow localhost http_access allow localnet http_access deny all icp_access allow localnet icp_access deny all http_port 3128 hierarchy_stoplist cgi-bin ? cache_dir ufs /var/spool/squid/cache1 1000 16 256 access_log /var/log/squid/access.log squid refresh_pattern ^ftp: 1440 20% 10080 refresh_pattern ^gopher: 1440 0% 1440 refresh_pattern -i (/cgi-bin/|\?) 0 0% 0 refresh_pattern (Release|Package(.gz)*)$ 0 20% 2880 refresh_pattern . 0 20% 4320 acl shoutcast rep_header X-HTTP09-First-Line ^ICY.[0-9] upgrade_http0.9 deny shoutcast acl apache rep_header Server ^Apache broken_vary_encoding allow apache extension_methods REPORT MERGE MKACTIVITY CHECKOUT cache_mgr webmaster cache_effective_user proxy cache_effective_group proxy hosts_file /etc/hosts coredump_dir /var/spool/squid access.log 1269243042.740 0 192.168.1.11 TCP_DENIED/400 2576 GET NONE:// - NONE/- text/html 00-firewall iptables -F iptables -t nat -F iptables -t mangle -F iptables -X echo 1 | tee /proc/sys/net/ipv4/ip_forward iptables -t nat -A POSTROUTING -j MASQUERADE iptables -t nat -A PREROUTING -p tcp --dport 80 -j REDIRECT --to-port 3128 networking auto lo iface lo inet loopback auto eth0 iface eth0 inet static address 142.104.109.179 netmask 255.255.224.0 gateway 142.104.127.254 auto eth1 iface eth1 inet static address 192.168.1.100 netmask 255.255.255.0

    Read the article

  • How to remove request blocking on apache reverse proxy after failure of backend before asking backen

    - by matnagel
    I am working on an apache2 reverse proxy vhost. When the server behind apache is down, the first request to apache shows the error page of course. But at subsequent requests it seems apache delays for some time before asking the backend server again. During all this time (which is short but in development I don't want a delay at all) only the apache error page is shown to the browser, although the backend server is already up. Where is this setting in apache, what is this behaviour, and how can I set the delay time to zero? Edit: I am not trying to change the timeout for a single request. I want to change the blocking time. It is my experience that apache blocks further requests for a certain time before asking a backend server again that has failed once. Edit2: This is what apache delivers: Service Temporarily Unavailable The server is temporarily unable to service your request due to maintenance downtime or capacity problems. Please try again later. Apache/2.2.8 (Ubuntu) PHP/5.2.4-2ubuntu5.7 with Suhosin-Patch proxy_html/3.0.0 Server at localhost Port 80 After hitting Ctrl-R in firefox for 60 seconds the page finally appears.

    Read the article

  • What's wrong with this HTTP POST request?

    - by bigboy
    I'm trying to fuzz a server using the Sulley fuzzing framework. I observe the following stream in Wireshark. The error talks about a problem with JSON parsing, however, when I try the same HTTP POST request using Google Chrome's Postman extension, it succeeds. Can anyone please explain what could be wrong about this HTTP POST request? The JSON seems valid. POST /restconf/config HTTP/1.1 Host: 127.0.0.1:8080 Accept: */* Content-Type: application/yang.data+json { "toaster:toaster" : { "toaster:toasterManufacturer" : "Geqq", "toaster:toasterModelNumber" : "asaxc", "toaster:toasterStatus" : "_." }} HTTP/1.1 400 Bad Request Server: Apache-Coyote/1.1 Content-Type: */* Transfer-Encoding: chunked Date: Sat, 07 Jun 2014 05:26:35 GMT Connection: close 152 <?xml version="1.0" encoding="UTF-8" standalone="no"?> <errors xmlns="urn:ietf:params:xml:ns:yang:ietf-restconf"> <error> <error-type>protocol</error-type> <error-tag>malformed-message</error-tag> <error-message>Error parsing input: Root element of Json has to be Object</error-message> </error> </errors> 0

    Read the article

< Previous Page | 16 17 18 19 20 21 22 23 24 25 26 27  | Next Page >