Search Results

Search found 2454 results on 99 pages for 'domains'.

Page 51/99 | < Previous Page | 47 48 49 50 51 52 53 54 55 56 57 58  | Next Page >

  • Apache VirtualHost Blockhole (Eats All Requests on All Ports on an IP)

    - by Synetech inc.
    I’m exhausted. I just spent the last two hours chasing a goose that I have been after on-and-off for the past year. Here is the goal, put as succinctly as possible. Step 1: HOSTS File: 127.0.0.5 NastyAdServer.com 127.0.0.5 xssServer.com 127.0.0.5 SQLInjector.com 127.0.0.5 PornAds.com 127.0.0.5 OtherBadSites.com … Step 2: Apache httpd.conf <VirtualHost 127.0.0.5:80> ServerName adkiller DocumentRoot adkiller RewriteEngine On RewriteRule (\.(gif|jpg|png|jpeg)$) /p.png [L] RewriteRule (.*) /ad.htm [L] </VirtualHost> So basically what happens is that the HOSTS file redirects designated domains to the localhost, but to a specific loopback IP address. Apache listens for any requests on this address and serves either a transparent pixel graphic, or else an empty HTML file. Thus, any page or graphic on any of the bad sites is replaced with nothing (in other words an ad/malware/porn/etc. blocker). This works great as is (and has been for me for years now). The problem is that these bad things are no longer limited to just HTTP traffic. For example: <script src="http://NastyAdServer.com:99"> or <iframe src="https://PornAds.com/ad.html"> or a Trojan using ftp://spammaster.com/[email protected];[email protected];[email protected] or an app “phoning home” with private info in a crafted ICMP packet by pinging CardStealer.ru:99 Handling HTTPS is a relatively minor bump. I can create a separate VirtualHost just like the one above, replacing port 80 with 443, and adding in SSL directives. This leaves the other ports to be dealt with. I tried using * for the port, but then I get overlap errors. I tried redirecting all request to the HTTPS server and visa-versa but neither worked; either the SSL requests wouldn’t redirect correctly or else the HTTP requests gave the You’re speaking plain HTTP to an SSL-enabled server port… error. Further, I cannot figure out a way to test if other ports are being successfully redirected (I could try using a browser, but what about FTP, ICMP, etc.?) I realize that I could just use a port-blocker (eg ProtoWall, PeerBlock, etc.), but there’s two issues with that. First, I am blocking domains with this method, not IP addresses, so to use a port-blocker, I would have to get each and every domain’s IP, and update theme frequently. Second, using this method, I can have Apache keep logs of all the ad/malware/spam/etc. requests for future analysis (my current AdKiller logs are already 466MB right now). I appreciate any help in successfully setting up an Apache VirtualHost blackhole. Thanks.

    Read the article

  • kerberos NTLM authentication

    - by rockbala
    Hi, Where can I check in Windows Server 2008 that Kereberos/NTLM is the authentication protocol used after the whole network is set up/installed. There is only 1 domain controller with AD services and is not affiliated with any other domains. Regards, Balaji S

    Read the article

  • Postfix configuration (relay access)

    - by jome
    I have just installed POSTFIX on a Debian box, I pointed the relay host config setting to an exchange server which will deliver the mail to external users. So what I am trying to do is telnet to the debian box and send an email to [email protected] which will then be past to the exchange server for delivery. The problem is I get the following: rcpt to:[email protected] 554 relay access denied I have seen a section in the config "mydestination" but i want the exchange server to decide which domains it will deliver for and not the POSTFIX server.

    Read the article

  • Set web browser (or other program) to use specific network

    - by Will
    my PC is connected to 3 different networks: 2 different LAN connections and a VPN connection. when i am connected to the vpn, i cannot browse to the LAN intranet as the DNS is all going through the vpn and it cannot see the domains on the LAN. Is there a way to set firefox/chrome/etc to use a specific connection, so that i don't have to disconnect from vpn to get to the other intranets?

    Read the article

  • ADFS 2.0 Server Prompts For Credentials When Using FQDN

    - by ncaudill
    We have an ADFS test enviroment set up, but we are running into issues with login prompts. If we browse to ADFS from Domain A we get a token sucessfully from ADFS, however when we browse from Domain B we are getting prompted for credentials. Domain A trusts Domain B but Domain B does not trust Domain A. The weird thing is, if we replace the full domain name with the server's IP address we can sucessfully get through from both domains. I feel like this should be a really simple solution, but we're stumped.

    Read the article

  • What would cause my SendMail server not to acknowledge receiving a TCP Sequence?

    - by Mike B
    My TCP/IP Stack knowledge is a little rusty so please bear with me.... I have a CentOS 5.7 server with SendMail and am having seeing intermittent timeout issues sending email (particularly larger email) to other remote domains. It doesn't happen with all attachments or recipient domains. Just some. After some extended troubleshooting, I think I've narrowed it down to TCP Sequences not being acknowledged. Here's a breakdown of the TCP session from a packet capture I collected directly on my MTA (fooMTA): Packet 1 - 11: Standard TCP handshake followed by initial SMTP conversation. No errors. Packet #12 Recipient MTA: TCP sequence 231. Ack 91. Packet #13 FooMTA: TCP sequence 91. Ack 305. Packet #14 FooMTA: TCP sequence 1115. Ack 305. Packet #15 Recipient MTA: TCP sequence 305. Ack 2495. Packet #16 FooMTA: TCP sequence 2495. Ack 305. Packet #17 FooMTA: TCP sequence 5255. Ack 305. Packet #18: Recipient MTA: TCP sequence 305. Ack 5255. Packet #19: FooMTA: TCP sequence 6635. Ack 305. Packet #20: FooMTA: TCP sequence 8015. Ack 305. Packet #21: Recipient MTA: TCP Sequence 305. Ack 8015. Packet #22: FooMTA: TCP Sequence 10775. Ack 305. Packet #23: FooMTA: TCP Sequence 13535. Ack 305. Packet #24: Recipient MTA: TCP sequence 305. Ack 10775 Packet #25: FooMTA: TCP Sequence 14915. Ack 305 It keeps going like this with my server still thinking it hasn’t received sequence 305… in response the remote side eventually retransmits its prior data thinking that it never arrived. Eventually the gap gets so large that no new data is sent and the remote MTA keeps retransmitting old stuff. This contributes to an exponential backoff and eventually the remote side gives up. What’s strange to me is that I see the “missing” TCP sequence (305 in this case) arriving back to my server (via a packet capture collected directly from fooMTA) So I don’t get why my server keeps asking for it. Could this be firewall related? What would be the next step in troubleshooting?

    Read the article

  • Best ASP.Net Host for Developers

    - by Tyler
    I would need it to allow me to host subdomains and multiple domains is a huge plus. Required: ASP.NET 2.0, 3.0, 3.5 Subdomain Hosting MS-SQL & MySQL Databases Want Multiple Domain Hosting ASP.NET 4.0 Ability to directly connect to MS SQL using SQL SMS So what do you have for me SF?

    Read the article

  • postfix default smtp route

    - by user42700
    hi, I am using postfix with mailscanner to scan emails. the smtp route(tranport_maps) works fine, but for domains on default route, i added this smtp:[mail.server.net] but it still gives access denied error, any ideas?

    Read the article

  • Troubleshooting Cpanel Email delays

    - by cpain
    I am running cpanel cPanel 11.25.0 on CentOS and have intermittment email delays (both coming and going) for a few specific domains. Emails from this server as well as major providers such as gmail, etc., do not suffer these delays. Any help with troubleshooting this issue? I am well aware this could be caused by something external but at this point getting someone else to fix this issue is out of the question.

    Read the article

  • Enabling Remote Desktop across different domain

    - by Sriram
    Have a system with Win8 within Domain A and remote has been enabled in that. Users from Domain A are able to login reomtely using admin credentials. how ever using the same credentials users from Domain B Domain C & Domain D are not able to login. Is there any setting that needs to be done so that users from other 3 domains are able to access the system either using the system admin account or using their login (added in remote desktop users account in the remote machine) Any suggestion will be helpful

    Read the article

  • How to Deploy an ASP.NET Web API- and Browser-based Application to a Production Environment

    - by user69508
    (Please forgive if this is posted in an incorrect forum. We didn’t know exactly where to post it.) We have an ASP.NET Web API single page application - a browser-based app running in IIS to serve up HTML5/CSS3/JavaScript, which talks to the ASP.NET Web API endpoint only to access a database and transfer JSON data. Everything is working great in our development environment - that is, we have one Visual Studio solution with an ASP.NET Web API project and two class library projects for data access. While development and testing on development boxes, using IIS Express to a localhost:port to run the site and access the Web API, everything is fine. Now we need to move it to a production environment (and we’re having problems - or just not understanding what needs to be done). The production environment is all internal (nothing will be exposed on the public Internet). There are two domains. One domain, the corporate domain, is where all users login normally. The other domain, the process domain, contains the SQL Server instance that our app and Web API will need to access. The IT staff wants to put a DMZ between the two domains to house the IIS app and shield the users on the corporate domain from having access into the process domain directly. So, I guess what they want is: corp domain (end users) <– firewall (open port 80) <– DMZ (web server running IIS) <– firewall (open port 80 or 1433????) <– process domain (IIS for Web API and SQL Server) We’re developers and don’t really understand all the networking aspects, so we’re wondering how to deploy our browser/Web API application in this scenario. Do we need to break up our application so that all the client code (HTML5/CSS3/JavaScript/images/etc.) is on the IIS server in the DMZ, while the Web API gets installed on the server in the process domain? Or, does the entire app (client code and Web API) stay together on the IIS server in the DMZ, which then somehow accesses the SQL Server instance to get data? From the IIS server and app in the DMZ, would you simply access the Web API on the server in the process domain by going to "http://server/appname/api/getitmes"? In the second firewall between the DMZ and the process domain, would you have to open port 1433 or just port 80 since the Web API is a HTTP endpoint? Or, is there some better way of deployment (i.e., how ASP.NET Web API single page applications written all in HTML5 and JavaScript supposed to be deployed to production environments?)? I’m sure there are other questions, but we’ll start with these. Thanks!!! (Note: the servers are Win2k8 R2, SQL Server 2k8 R2, and IIS 7.5.)

    Read the article

  • Creating a fallback error page for nginx when root directory does not exist

    - by Ruirize
    I have set up an any-domain config on my nginx server - to reduce the amount of work needed when I open a new site/domain. This config allows me to simply create a folder in /usr/share/nginx/sites/ with the name of the domain/subdomain and then it just works.™ server { # Catch all domains starting with only "www." and boot them to non "www." domain. listen 80; server_name ~^www\.(.*)$; return 301 $scheme://$1$request_uri; } server { # Catch all domains that do not start with "www." listen 80; server_name ~^(?!www\.).+; client_max_body_size 20M; # Send all requests to the appropriate host root /usr/share/nginx/sites/$host; index index.html index.htm index.php; location / { try_files $uri $uri/ =404; } recursive_error_pages on; error_page 400 /errorpages/error.php?e=400&u=$uri&h=$host&s=$scheme; error_page 401 /errorpages/error.php?e=401&u=$uri&h=$host&s=$scheme; error_page 403 /errorpages/error.php?e=403&u=$uri&h=$host&s=$scheme; error_page 404 /errorpages/error.php?e=404&u=$uri&h=$host&s=$scheme; error_page 418 /errorpages/error.php?e=418&u=$uri&h=$host&s=$scheme; error_page 500 /errorpages/error.php?e=500&u=$uri&h=$host&s=$scheme; error_page 501 /errorpages/error.php?e=501&u=$uri&h=$host&s=$scheme; error_page 503 /errorpages/error.php?e=503&u=$uri&h=$host&s=$scheme; error_page 504 /errorpages/error.php?e=504&u=$uri&h=$host&s=$scheme; location ~ \.(php|html) { include /etc/nginx/fastcgi_params; fastcgi_pass 127.0.0.1:9000; fastcgi_intercept_errors on; } } However there is one issue that I'd like to resolve, and that is when a domain that doesn't have a folder in the sites directory, nginx throws an internal 500 error page because it cannot redirect to /errorpages/error.php as it doesn't exist. How can I create a fallback error page that will catch these failed requests?

    Read the article

  • Exim To Include Internal Email When Sending

    - by user219951
    I wonder if you can help me out... at the moment were using exim to relay our emails to an smtp_relay. The only problem is if the user exists and has an email address it won't exit the server and send them mail using our relay. I need it to go to the relay even if it's an internal email address. Hope that make sense, this is how we have got our relay setup at the moment smart_route: driver = manualroute domains = !+local_domains transport = remote_smtp_smtpauth route_list = * smtp.sendgrid.net:587 Thanks

    Read the article

  • WHMCS on main website [closed]

    - by Miskone
    I ordered a Dedicated Root Server on Hetzner. I have WHM and cPanel and now I want to make front site for histing company where users can control their accounts, nameservers of registered domain, register new domains and new hosting accounts. My question is: Do I need to buy WHMCS for that or something else. I need to put client zone on front site and I want to find the best solution for that. Sorry if this is not topic for this site, but I just don't know where can I ask this question :(

    Read the article

  • Does hosting multiple sites on one server hurt your SEO?

    - by MarathonStudios
    I have a handful of (content unrelated) sites with decent PRs and I'm considering hosting them all on the same server. I've heard that if you do this, internal linking between two seperate domains on that server may be seen as less "valid" by Google in PageRank terms (since you obviously own both of the sites as they share an IP address). Anyone have any experience in this? I'd love to save some hosting cash by consolidating, but not at the expense of losing the ability to link my sites together powerfully.

    Read the article

  • virutal users in postfix

    - by facha
    Hello everyone, I'd like to setup postfix as an MTA for 2 domains, about 10 users each. I'm a bit confused with the documentation. Postfix has so many options for mail delivery. I'd like to have a simple file where I would store all the email addresses (and possibly their corresponding mailboxes). I wouldn't like neither system's passswd file nor mysql to be used for storing email users. What parts of documentation I should be focused on?

    Read the article

  • Custom settings with PHP on Nginx

    - by miki
    I have multiple websites setup using Nginx and Apache, but when I try to add a vhost based PHP directive using fastcgi_param PHP_VALUE , the value get added to all of my vhosts. Checking the value of the PHP parameter from PHP-CLI remain the actual one for whole server. e.g. I used fastcgi_param PHP_VALUE "memory_limit=512M" for a domain based nginx config, but it will be propagated to all domains on server. The 'php -i | grep memory_limit' show 128M though Not sure what I am missing

    Read the article

  • Exchange Not Delivering Email

    - by wolfvilleian
    I'm having an issue where my Exchange Edge Transport server receives mail (found in logs) and then it vanishes, never ending up in the users mailbox, I have a edge subscription setup between it and the main Exchange server, how can I go about tracing the message to figure out what is broken? I also have found records of the message in the logs on the main Exchange server. Thanks a ton for any help Edit: If I change port 25 on my main router to point to the main exchange server as opposed to the Edge Transport, email comes through fine form external domains and delivered in the correct mailbox

    Read the article

  • How to block a specific site's directory in windows?

    - by Creedy
    How can I block a specific directory of a website as e.g. http://example.com/someSite unfortunately the hosts file is not an option since you can only block whole domains there and any "/" just destroys these rules. This is just for my personal protection against visiting some sites too often, while i still have to be able to get to the other sites of that domain (as e.g. example.com/someOtherSite) Would be great if someone knows a solution regarding this topic.

    Read the article

  • Tips for Domain Name Management?

    - by bofe
    Expired domain names = downtime for websites. Downtime = bad. How does your organization make sure domain names have been renewed? I believe ICANN requires registrars to give a notice at 60 days and 30 days, but these can easily get ignored -- especially with a large amount of domains. Does your solution work for a large amount of domain names? ( 100) Is it registrar specific?

    Read the article

< Previous Page | 47 48 49 50 51 52 53 54 55 56 57 58  | Next Page >