Search Results

Search found 7082 results on 284 pages for 'trs 80'.

Page 43/284 | < Previous Page | 39 40 41 42 43 44 45 46 47 48 49 50  | Next Page >

  • Running Tomcat 7 and Apache 2 on the same server

    - by Thorn
    Part of my site needs to run over HTTPS and I'm creating a sub-domain for that part. I have apache httpd 2 AND Tomcat 7 running on the same server with the same IP, Apache is on port 80 of course, while Tomcat is running on port 8080. Right now I am doing domain forwarding for requests that need to run off tomcat. For example, mathteamhosting.com/mathApp can forward to mathteamhosting.com:8080/mathApp. I would like to have Tomcat handle the https requests for that subdomain. I don't think this forwarding technique can work in this case. How do I set that up so that Tomcat receives the requests on port 443 while apache handles port 80. To be more specific: http://proctinator.com == request goes to Apache web server https://private.proctinator.com == request goes to Apache web server

    Read the article

  • output redirection not happening completely

    - by Gaurav
    I am running the following command on ubuntu 12.04 bash shell: hping3 -S -p 80 -c 30 www.google.com >> outputfile.txt But the problem is that this leaves some lines in the end and doesn't copy in the file. I even tried hping3 -S -p 80 -c 30 www.google.com >> outputfile.txt 2>&1 It also gives same problem. However this works perfect: ping -c 30 www.google.com >> outputfile.txt Any idea about what I might be missing?

    Read the article

  • What does *:* in netstat output stands for?

    - by chello
    While executing the command /usr/sbin/lsof -l -i -P -n under root user, I am getting this output. COMMAND PID USER FD TYPE DEVICE SIZE/OFF NODE NAME ... httpd 9164 70 3u IPv4 0x2f70270 0t0 TCP 127.0.0.1:9010 (LISTEN) httpd 9164 70 4u IPv6 0x25af4bc 0t0 TCP *:80 (LISTEN) httpd 9164 70 5u IPv4 0x3149e64 0t0 TCP *:* (CLOSED) httpd 9180 70 3u IPv4 0x2f70270 0t0 TCP 127.0.0.1:9010 (LISTEN) httpd 9180 70 4u IPv6 0x25af4bc 0t0 TCP *:80 (LISTEN) httpd 9180 70 5u IPv4 0x3149e64 0t0 TCP *:* (CLOSED) Please let me know what does *:* stands for? I am interested to know both the ipaddress and port fields. Also what does (CLOSED) mean here?

    Read the article

  • SSL on app - nginx web server

    - by Adam
    I am running an nginx web server where I redirect all http requests to https (with a self signed cert). Here is how I REDIRECT all http requests to https in the nginx config file: server { listen 80 default_server; listen [::]:80 default_server ipv6only=on; server_name my.server.ip; return 301 https://$server_name$request_uri; Problem is - I cannot seem to do so for an app running on a port. Example: http://my.server.ip:1234 does not redirect to https://my.server.ip:1234 ir works fine on all other urls like http://my.server.ip/temp etc. How can I modify the nginx config file to force that app url through ssl?

    Read the article

  • Outlook '10 hangs often, IMAP sync

    - by user23150
    We have 3 employees using IMAP to sync with their Desktop and Android Phones. Two are using 70% of their accounts storage, another is using 80%. They all have similar counts on folder structure, etc. The employee with the 80% storage is constantly having Outlook freeze on them for up to 5 minutes at a time. I realize this is Outlook connecting and doing activity on the server, but no one else has this problem. In fact, one user with the 70% storage used uses a very slow laptop, and doesn't have freezing issues. The network is the same, the settings are the same - at a loss how to proceed? Obviously "Outlook is a crummy IMAP client" doesn't help management...

    Read the article

  • Apache httpd VirtualHost config - multiple sites

    - by DaFoot
    [Advised to post here from StackExchange] I have a site to work on, because of the way the URLs are built the application seems to have been created on the assumption that it will be at the server root (only app). On my dev server I have other projects and up to now a simple symlink has been working for me, but that's not the case now because this new app wants to sit at the route and process all URLs arriving on :80. Hopefully this snippet from httpd.conf will help explain what I'm trying to acheive: # default for any not matched elsewhere <VirtualHost *:80> ServerName localhost DocumentRoot /var/www/html/newproject </VirtualHost> # now try to pick out specific URLs <VirtualHost localhost/webdev> DocumentRoot /var/www/html/existingProject ServerName localhost/project </VirtualHost> Also need to be able to get same affect from wherever I'm accessing the httpd instance from. Hope that makes sense.

    Read the article

  • ssh can't connect after server ip changed

    - by Kery
    I have a server with ubuntu installed. After I change the network configuration and restart server, ssh client can't connect server any more. But in the server I can use ssh client to connect itself and the netstat command shows that sshd is listening port 22. And in my computer (win7) ping command is OK to server's new IP. The configuration in /etc/network/interfaces is: auto eth0 iface eth0 inet static address 10.80.x.x netmask 255.255.255.0 gateway 10.80.x.1 I'm very confused about this. Hope somebody can give me some idea. Thank you in advance!!!

    Read the article

  • Lighttpd based server issues crop up when port forwarding

    - by michael
    I have four host computers running lighttpd webservers. they are sitting behind a hspa modem, which each occupying a http port between [81 - 84]. 80 is taken by the modem itself. The port forwarding is setup correctly, however, only a portion of any webpage I request from any of the hosts comes through (they all fails after %20 of the page). If I put the host on port 81 into the dmz, it serves pages fine. The others do not respond to the dmz treatment. Is it possible the web content on the hosts somehow require ports aside from their respective http port? Or is it possible that even though the server.port in the lighttpd_ssl.conf file is set, the individual hosts are still expecting to serve on port 80? I am not familiar with lighttpd, nor did i set them up. they are running on video encoders i purchased. I can grab any files from them required for further information on the problem.

    Read the article

  • Starting Apache through XAMPP is not working

    - by nogamo
    I've checked the error logs for this particular problem and it seems no access can be made to port 80. (OS 10013)An attempt was made to access a socket in a way forbidden by its access permissions. : make_sock: could not bind to address [::]:80 Turning the Windows firewall off didn't work, so I'm a bit lost. All of this is running on my laptop so I don't have my own "server" per se but I was under the impression that this would work without the need for one independent of my own personal computer. How would I solve this?

    Read the article

  • Proxy between data centers [closed]

    - by dstarh
    Possible Duplicate: Can IIS be configure to forward request to another web server? We are switching data centers (actually datac-center to EC2 to be specific) and some customers have not yet made DNS changes to point the domains to the new load balancers. We are thinking of leaving the existing servers up and just using a proxy server to forward the request to the new load balancer. Can anyone recommend a good proxy server for doing this I've got squid installed but it seems it's fairly easy to just have a fairly wide open proxy server and we don't want this. I want all requests coming in on port 80 to be sent to port 80 at a specific domain (elastic load balancer) the data center env is windows 2k3 and the EC2 env will all be linux but the ec2 env should be irrelevant

    Read the article

  • ZendServer uninstall and xampp install ports conflict

    - by BlackFire27
    I uninstalled zend, cleaned it from the registry (it doesnt exist there).. when I go to localhost , I can see its favicon. Also my xampp port is blocked by the previous installation of zend that uses the port 80.. so I swapped xampps port to 8080.. it works..but I would rather to use port of 80.. is there a way to achieve so? perhaps changing zends port to something else? or how can I find who is listening to the same port.. By what I can see is that zend is still using the localhost..how i sthat possible

    Read the article

  • EC2: map multiple applications to different domains

    - by EsseTi
    i'm playing with EC2 and i've been able to create my instance that has a django appliacation on port 80, and a tomcat on 8080. now, with elastic IP i can manage to redirect my domain to django application. now i would like to map subdomains to each tomact applications. for example django app (ec2...:80) --> mydomain.com tomcat (ec2...:8080) --> tomcat.mydomain.com webbapp1 (ec2...:8080/webapp1/) --> webapp1.mydomain.com is this possible with the free account? ciao

    Read the article

  • Installing CURL on Ubuntu Karmic

    - by Racertim
    Trying to get this up and running: https://github.com/cloudnull/massupload I have everything except CURL installed and when I attempt to, it fails with the following: Reading package lists... Done Building dependency tree Reading state information... Done The following NEW packages will be installed: curl 0 upgraded, 1 newly installed, 0 to remove and 3 not upgraded. Need to get 196kB of archives. After this operation, 311kB of additional disk space will be used. WARNING: The following packages cannot be authenticated! curl Install these packages without verification [y/N]? y Err http://us.archive.ubuntu.com karmic/main curl 7.19.5-1ubuntu2 404 Not Found [IP: 91.189.91.13 80] Failed to fetch http://us.archive.ubuntu.com/ubuntu/pool/main/c/curl/curl_7.19.5-1ubuntu2_i386.deb 404 Not Found [IP: 91.189.91.13 80] Thank you!

    Read the article

  • problem with accessing a php page

    - by EquinoX
    So I have a info.php page which is located on the folder /var/www/nginx-default, however when I go to my ip address/info.php, it always redirects me to this site: http://www.iana.org/domains/example/ is this because I have a virtual host that I called example? Here is my config for the example website: server { listen 80; server_name www.example.com; rewrite ^/(.*) http://example.com/$1 permanent; } server { listen 80; server_name example.com; access_log /var/www/example.com/logs/access.log; error_log /var/www/example.com/logs/error.log; location / { root /var/www/example.com/public/; index index.html; } } The way I access this site is by changing my /var/hosts in my macbook so that example.com is mapped to my server IP address... however now when I do xxx.xxx.xxx.xxx/info.php.. it redirects me to that site I posted above

    Read the article

  • wamp remote access

    - by user1589779
    I have a wamp server (Version 2.2 on Windows 7) running on my computer that works perfectly well for localhost access. But now, I need to grant access to a collegue on the same subnetwork. How to achieve that? My collegue receives a 403 error from the browser. I have 2 vhost : <VirtualHost *:80> ServerName localhost DocumentRoot "D:/Workspace" </VirtualHost> <VirtualHost *:80> ServerName oscar DocumentRoot "D:/Workspace/SMACS3/web/htdocs" </VirtualHost>

    Read the article

  • HAproxy to web host sub directory?

    - by daemonza
    Hi for reasons outside my control, I need to load balance two servers, that run a non-virtual host enabled app on IIS. Normally in HAProxy I would load balance servers(apache, tomcat, etc) like this : acl is_www_example_com hdr_end(host) -i www.example.com use_backend www_example_com if is_www_example_com backend www_example_com balance roundrobin cookie SERVERID insert nocache indirect option httpchk HEAD / HTTP/1.0 option httpclose option forwardfor server node1 192.168.1.1:80 cookie node1 server node1 192.168.1.2:80 cookie node1 Which will route to the node 1 and node 2 server and serve up the virtual host site. if I need to route to www.example.com/application/data How would I be able to do it, with the above example, if at all even possible?

    Read the article

  • domain works randomly

    - by mthenw
    Hi, I have problem with configuring bind9 on debian (lenny) server. Generaly speaking everything is working ok but sometimes I get 404 on few domains (eg. 4stopnie.com but after few refreshes in browser site loads) or I can't validate site with validator.w3.org (error '500 Can't connect to 4stopnie.com:80 (Bad hostname '4stopnie.com')'). Domains were moved from other server. After moving I changed serial number in zone file. $ttl 600 @ IN SOA ns.wpoznaniu.info. xxx.4stopnie.com. ( 2011011601 3600 600 86400 600) @ IN NS ns.wpoznaniu.info. @ IN A 80.82.21.196 www IN CNAME @

    Read the article

  • How to make your apache application accessible within network

    - by guest
    I have a Windows XP machine where I have installed WAMP and made a PHP based web application. I can access the web application from within this machine by using the browser and pointing to: http://localhost/myApp/ --- and the page loads fine. Now I want this site (http://localhost/myApp) to be accessible to all machines within the network (and may be later, to the general public as well). I am quite new to this, how do I make my site accessible to all machines within the network and to the general public in the internet? I tried modifying the httpd.conf file in Apache (WAMP) by changing Listen 80 to Listen 10.10.10.10:80 (where I replaced 10.10.10.10 with the actual IP of this windows xp machine). I also tried "Put Online" feature in WAMP. None seem to work though. How do I make it accessible?

    Read the article

  • Security against IP spoofing [on hold]

    - by user1369975
    I am pursuing a college project, in which I am running three fake services on three ports to protect the main service (say running at port 80). The concept is that if the user is malicious, he'll try to bring the services down and access the fake services. These ports adopt a blocking process of a connection request and record the IP and port of the client. These are logged and aren't granted access on service on port 80. But what to do if the client spoofs his IP? How can I modify my system?

    Read the article

  • Easiest method of creating multiple Word documents with incremental number?

    - by DakotaDusty
    I need to create about 80 Word documents that are identical except for a numerical identifier in the document. The identifier is the same as the document filename, eg. the document named "SN-100.doc" must have the string "SN-100" inside the document text. Each unique document will reside in a folder location given by its unique numerical identifier.For example, the file directory hierarchy will look something like this: /SN001/SN-1.doc /SN002/SN-2.doc . . . /SN080/SN-80.doc What is the easiest and fastest method of doing this?

    Read the article

  • radvd won't accept non-/64 subnets

    - by Dolda2000
    I'm trying to set up radvd on a 6RD configuration (on Linux), where I have a /64 subnet, and I'm trying to use it on two distinct physical networks, so I'm trying to give each of them a /80 subnet. However, when I start radvd on these interfaces, it says this: radvd[3987]: prefix length should be 64 for int (int being the name of one of the interfaces.) I know that stateless autoconfiguration doesn't work on subnets that are larger than /64, like a /48 subnet, but AFAIK it's supposed to work on anything that is smaller than /64, so /80 shouldn't be a problem. The final effect, now, is that radvd simply advertises the wrong subnet prefix, effectively "removing" the first 16 bits of it. What is this? Am I all wrong on that, is radvd buggy, or is there some configuration option that I'm missing?

    Read the article

  • apache 2.2: how can I set up a VirtualHost inside the RootDirectory?

    - by redraw
    I want to set up a VirtualHost inside the RootDirectory. For example, My project is in C:/myproject and I want to access with http://localhost/myproject EDIT: I've made an alias inside the httpd-vhosts.conf, however I don't have permissions. <VirtualHost *:80> DocumentRoot "C:/apache-2.2/htdocs" ServerName localhost Alias /test "D:\arbol\documentos\test" </VirtualHost> Is this code below the proper way to give permissions? <VirtualHost *:80> DocumentRoot "C:/apache-2.2/htdocs" ServerName localhost Alias /test "D:\arbol\documentos\test" <Directory "D:\arbol\documentos\test"> allow from all order allow,deny AllowOverride All </Directory> </VirtualHost>

    Read the article

  • Need help with setting up comet code

    - by Saif Bechan
    Does anyone know off a way or maybe think its possible to connect Node.js with Nginx http push module to maintain a persistent connection between client and browser. I am new to comet so just don't understand the publishing etc maybe someone can help me with this. What i have set up so far is the following. I downloaded the jQuery.comet plugin and set up the following basic code: Client JavaScript <script type="text/javascript"> function updateFeed(data) { $('#time').text(data); } function catchAll(data, type) { console.log(data); console.log(type); } $.comet.connect('/broadcast/sub?channel=getIt'); $.comet.bind(updateFeed, 'feed'); $.comet.bind(catchAll); $('#kill-button').click(function() { $.comet.unbind(updateFeed, 'feed'); }); </script> What I can understand from this is that the client will keep on listening to the url followed by /broadcast/sub=getIt. When there is a message it will fire updateFeed. Pretty basic and understandable IMO. Nginx http push module config default_type application/octet-stream; sendfile on; keepalive_timeout 65; push_authorized_channels_only off; server { listen 80; location /broadcast { location = /broadcast/sub { set $push_channel_id $arg_channel; push_subscriber; push_subscriber_concurrency broadcast; push_channel_group broadcast; } location = /broadcast/pub { set $push_channel_id $arg_channel; push_publisher; push_min_message_buffer_length 5; push_max_message_buffer_length 20; push_message_timeout 5s; push_channel_group broadcast; } } } Ok now this tells nginx to listen at port 80 for any calls to /broadcast/sub and it will give back any responses sent to /broadcast/pub. Pretty basic also. This part is not so hard to understand, and is well documented over the internet. Most of the time there is a ruby or a php file behind this that does the broadcasting. My idea is to have node.js broadcasting /broadcast/pub. I think this will let me have persistent streaming data from the server to the client without breaking the connection. I tried the long-polling approach with looping the request but I think this will be more efficient. Or is this not going to work. Node.js file Now to create the Node.js i'm lost. First off all I don't know how to have node.js to work in this way. The setup I used for long polling is as follows: var sys = require('sys'), http = require('http'); http.createServer(function (req, res) { res.writeHead(200, {'Content-Type': 'text/html'}); res.write(new Date()); res.close(); seTimeout('',1000); }).listen(8000); This listens to port 8000 and just writes on the response variable. For long polling my nginx.config looked something like this: server { listen 80; server_name _; location / { proxy_pass http://mydomain.com:8080$request_uri; include /etc/nginx/proxy.conf; } } This just redirected the port 80 to 8000 and this worked fine. Does anyone have an idea on how to have Node.js act in a way Comet understands it. Would be really nice and you will help me out a lot. Recources used An example where this is done with ruby instead of Node.js jQuery.comet Nginx HTTP push module homepage Faye: a Comet client and server for Node.js and Rack To use faye I have to install the comet client, but I want to use the one supplied with Nginx. Thats why I don't just use faye. The one nginx uses is much more optimzed. extra Persistant connections Going evented with Node.js

    Read the article

< Previous Page | 39 40 41 42 43 44 45 46 47 48 49 50  | Next Page >