Search Results

Search found 16001 results on 641 pages for 'convention over configuration'.

Page 274/641 | < Previous Page | 270 271 272 273 274 275 276 277 278 279 280 281  | Next Page >

  • Echo 404 directly from nginx to improve performance

    - by user64204
    I am in charge of production servers serving static content for a website. Those servers are constantly being crawled by bots looking for potential exploits (which isn't that much of a problem security-wise because no application can be reached behind the web server) but generates thousands of 404 per day, sometimes per hour. I am looking into ways of blocking those requests but it's tricky (you want to make sure you don't block legitimate traffic and these bots are becoming more and more clever at looking like they're legit) and is going to take me a while to find an acceptable solution. In the meantime I would like to reduce the performance impact of serving those 404 pages. Indeed we're using nginx which by default is configured to serve it's 404 page from the disk (This can be changed using the error_page directive but in the end the 404 will either have to be served from disk or from another external source (e.g. upstream application which would be worst)) which isn't ideal. I ran a test with ab on my local machine with a basic configuration: in one case I echo a message directly from nginx so the disk isn't touched at all, in the other case I hit a missing page and nginx serves its 404 from disk. server { # [...] the default nginx stuff location / { } location /this_page_exists { echo "this page was found"; } } Here are the test results (my laptop has Intel(R) Core(TM) i7-2670QM + SSD in case you're wondering why they are so high): $ ab -n 500000 -c 1000 http://localhost/this_page_exists Requests per second: 25609.16 [#/sec] (mean) $ ab -n 500000 -c 1000 http://localhost/this_page_doesnt_exists Requests per second: 22905.72 [#/sec] (mean) As you can see, returning a value with echo is 11% ((25609-22905)÷22905×100) faster than serving the 404 page from disk. Accordingly I would like to echo a simple 404 Page not Found string from nginx. I tried many things so far but they all failed, essentially the idea was this: location / { try_files $uri @not_found; } location @not_found { echo "404 - Page not found"; } The problem is that as soon as the echo directive is used, the http response code is set to 200. I tried changing that by doing error_page 200 = 400 but that breaks the configuration. How can I serve a 404 page directly from nginx? (without hacking the source which may be might next step)

    Read the article

  • Is it possible to monitor an Asterisk ConfBridge from a URL / Browser?

    - by Lorin S.
    I have an Asterisk server set up with minimal configuration, including the following confbridge definition / extension: *confbridge.conf* [testbridge] type=bridge video_mode=follow_talker max_members=20 mixing_interval=10 internal_sample_rate=auto record_conference=yes *extension.conf* exten => 6100,1,Answer() same => n,Set(CONFBRIDGE(user,admin)=yes) same => n,Set(CONFBRIDGE(user,marked)=yes) same => n,ConfBridge("Ad-hoc",testbridge,default_user,sample_user_menu) Is it possible to monitor the video / audio of the conference without joining via a client?

    Read the article

  • multi-screen switcher

    - by João Melo
    I have a windows 7 machine with vmware and 2 monitors. The first one, with vmware on fullscreen, and the second one as the main screen with windows 7 with the browser and some applications, is there any way to toggle the screens configuration with a hotkey so it switches the main screen to the left and all applications accordingly (Vmware on full screen on the right monitor, and all applications on the left)

    Read the article

  • Help me hosting Apache server on a LAN connected computers?

    - by akhilesh
    i have build a JSP project,but i really want that other computer connected to my computer can access by website.i have never done it before Please help me.Now my server can be accessed using http://localhost:8080 on my local machine.what are the configuration i need to do please tell me.please post some link or step by step help.

    Read the article

  • Windows 7: Edit group policy from command line

    - by user234461
    I'm writing an installer and need to change all users' wallpaper. I can do this from the group policy editor GUI, but need to do so from my installer. I can't just edit the registry as it gets reset by a GPO on login. How do I apply the relevant administrative template via programatically (preferably cmd.exe or via the registry)? (for interest, it's User Configuration Administrative Templates Desktop Desktop Desktop Wallpaper [sic] Any help would be appreciated. Thanks!

    Read the article

  • Apache returns 403 Forbidden for alternative port vhost

    - by Wesley
    I'm having an issue getting vhosts to work on Apache 2.2, Debian 6. I have two VirtualHosts, one on port 80 and one on port 8888. The port 80 one has been created automatically by DirectAdmin, the 8888 is a custom one. It's configuration is as follows. <VirtualHost *:8888 > DocumentRoot /home/user/public_html/development ServerName www.myserver.nl ServerAlias myserver.nl <Directory "/home/user/public_html/development"> Options +Indexes +FollowSymLinks +MultiViews AllowOverride All Order Allow,deny Allow from all </Directory> </VirtualHost> Of course I also have a NameVirtualHost *:8888 The port 80 DocumentRoot is /home/user/public_html/production, which is perfectly accessible and works like a charm. The port 8888 docroot of /home/user/public_html/development is 403 forbidden though. I have compared the permissions for both folders. They seem fine to me. drwxr-xr-x 2 root root 4096 Aug 17 16:14 development drwxr-xr-x 4 root root 4096 Aug 18 04:29 production Also, the index.php file which is supposed to display when accessing through port 8888, located in /development/: -rwxr-xr-x 1 root root 41 Aug 17 16:14 index.html I have looked at my error_log and found many of the following entries, only being added to the log file when accessing through port 8888. [Sat Aug 18 04:35:09 2012] [error] [client 27.32.156.232] Symbolic link not allowed or link target not accessible: /home/user/public_html /home/user/public_html is a symbolic link that refers to /home/user/domains/mydomain/public_html. The symbolic link has the following permissions: lrwxrwxrwx 1 admin admin 29 Aug 17 15:56 public_html -> ./domains/mydomain/public_html I'm at a loss. It seems that everything is readable or executable. I've set the Directory to FollowSymLinks in the httpd.conf file, but that doesn't seem to make a difference. If I change that directory tag to <Directory "/home/admin/public_html"> (so it has FollowSymLinks on that as well) it still does not work. Any help is greatly appreciated. If I need to post more information, let me know. I'm pretty much a beginner at this stuff. .. .. UPDATE: I ended up changing the configuration to directly go to the actual path of the files, avoiding the public_html symlink altogether. That worked. Thanks for the suggestions folks. DocumentRoot /home/user/domains/mydomain/public_html/development instead of DocumentRoot /home/user/public_html/development

    Read the article

  • QEMU & libvirt: Dual screen VM?

    - by thecapsaicinkid
    I'm running a Windows 7 guest virtualised under QEMU and using libvirt on a Fedora 17 host, does anyone know if it's possible to create 2 video 'cards' (or a single dual-head) and a spice server for each and connect multiple spice clients to simulate a dual screen VM? Looking through the VMs xml configuration file I can't see a way to associate elements (eg. Spice servers) to elements (in this case, a qxl card). I'm sure qxl doesn't support the dual-head option but are multiple video elements possible?

    Read the article

  • Group Policy is not being applied from Server 2003 to win7 client

    - by John Hoge
    Hi, I'm experimenting with Group Policy settings. My DC is running Server 2003, and the client I am using for this test is running Win7. I've restarted the client a few times, and tried running gpupdate/force for good measure. This machine is in it's own OU with a group policy applied to change one setting, Computer Configuration/Administrative Templates/Network/Offline Files. When I run MMC and look at Local Computer Policy on the client this setting shows up as "not configured". Thanks, John

    Read the article

  • Secure linux distro that sends traffic through Tor

    - by gravestone
    I'm looking for a live Linux distro that is secure and preserves my anonymity online. For what I need, Tor seems to do the job – but such software requires configuration that someone who is not knowledgeable in how it works might find difficult. I was looking at 'Lightweight portable security' but it does not specify whether it sends all traffic through Tor. Does anyone know more information about it and/or any other such live Linux distro?

    Read the article

  • Git multiple config files

    - by tig
    As it is told in git-config manual, I have three files to specify git repo configuration: $GIT_DIR/config, ~/.gitconfig and $(prefix)/etc/gitconfig. Is it possible to make git read another config file (for example .gitconfig in root of the repo) after reading $GIT_DIR/config? I want it as I created git command aliases in .git/config and I would prefer to store them in repo, so I can restore them when cloning and for example.

    Read the article

  • Printer redirection on server 2003

    - by user137841
    On windows server 2003 when one user connects to the server via RDP the default printer of the server for her profile does not change to the redirected printer of the session. This only happens with the one user all the other users default printers defaults to their session printer automatically. I tried the following solution but there was no \Terminal Server\Printer Redirection in gpedit.msc http://technet.microsoft.com/en-us/library/cc731963(v=ws.10).aspx Computer Configuration\Policies\Administrative Templates\Windows Components\Terminal Services\Terminal Server\Printer Redirection Is there a different place to check the Printer Redirection?

    Read the article

  • centos freezes with this error kernel: ata1.00: exception Emask 0x0 SAct 0x7fffffff SErr 0x0 action 0x0

    - by lakshman
    0 down vote favorite share [fb] share [tw] I am using centOs 5.5 version with raid 1 configuration the server freezes and goes to non response . the only thing i found on messages file is kernel: ata1.00: exception Emask 0x0 SAct 0x7fffffff SErr 0x0 action 0x0 The server is built recently Please let us know what is the problem the hard disk details are Model Number: ST500NM0011 Serial Number: Z1M02LT7 Firmware Revision: SN02

    Read the article

  • Can't add a .ldif to OpenLDAP, recent version (no slapd.conf)

    - by Biganon
    I'm new to LDAP and I'm trying to add the mmc.ldif and mail.ldif files that come with Mandriva Directory Server to my LDAP configuration, using the command : ldapadd -x -W -D "cn=admin,dc=biganon,dc=com" -f schema/mmc.ldif I then give the admin password I've set during slapd installation but get this error : ldap_bind: Invalid credentials (49) I have no slapd.conf file. I'm on Debian 6.0.1, OpenLDAP 2.4.23 Thank you

    Read the article

  • Change drive letter of DVD drive

    - by NickC
    I am trying to use a Drive Maps GPO on Server 2012 - User Configuration - Preferences - Windows Settings - Drive Maps. Problem is I want to map a drive to D: but this is automatically assigned to the DVD drive. Question is can I prevent D: being used as the next drive letter when Windows 8 boots up? I seem to recall in early version of Windows, or maybe it was Novell, a first_drive environment variable for this purpose.

    Read the article

  • Data replication between two web nodes

    - by HTF
    I have Wordpress installation running on two web servers (Nginx). There is unidirectional synchronization from server A to server B and I'm using lsyncd for this purpose. with his configuration I have to add blog posts from the first web server so the data is replicated to the second one - how I can force access to Wordpress back-end only from the first web server? Please note that both servers have the same domain for Wordpress. Regards

    Read the article

  • Can Gitosis enforce correct user name/email?

    - by koumes21
    Gitosis is able to authenticate users based on public/private key pair. It is able to find out which user is currently committing. However, the user name and email is taken from the client's Git configuration ('git config user.name' etc.), which can be set to arbitrary values. Is there any way to associate user names and emails with their public keys and then make Gitosis uses these names and emails as the name and email of the committer?

    Read the article

  • Configure exim in debian 6

    - by blakcaps
    I am trying to configure exim with gmail in my debian 6 system as per this tutorial http://www.manu-j.com/blog/wordpress-exim4-ubuntu-gmail-smtp/75/.After configuring, When i run update-exim4.conf i am getting this message, Exim configuration error: two client authenticators (gmail_login and login) have the same public name (LOGIN) Invalid new configfile /var/lib/exim4/config.autogenerated.tmp, not installing /var/lib/exim4/config.autogenerated.tmp to /var/lib/exim4/config.autogenerated Any pointers to solve this?

    Read the article

  • Recompile PHP5.3 with mysqlnd on Debian Squeeze?

    - by TheRebel
    My current PHP5.3 installation was compiled with libmysql by default. To run certain mysqli functions, I need the mysqlnd library, but I can't seem to find a step by step guide giving proper description how to recompile PHP 5.3 on Debian 6 with custom settings. All I found was the configuration parameter on the PHP.net manual (http://php.net/manual/en/mysqlinfo.library.choosing.php) saying that this is a compile time decision: $ ./configure --with-mysqli=mysqlnd --with-pdo-mysql=mysqlnd --with-mysql=mysqlnd Any ideas how to do it? Thanks!

    Read the article

  • Obtaining a DHCP IP address if static address is not available

    - by user142485
    I have seen some questions that are similar to my problem but not a good solution. Ex. User has a laptop and has a static IP address at the office. They also take the laptop off site where they also need internet. Is there a way to configure their computer so that it will obtain an IP address automatically when the configured static address is not available (without making the user go into network configuration and changing to obtain automatically)?

    Read the article

  • Nginx not working properly on subdomains [SOLVED]

    - by javipas
    I've been trying to setup a Sugar CRM instance. I've got a domain that has its main site on a server (www.domain.com) and I've created a subdomain (sugar.domain.com), but I wnat this subdomain to be hosted on another server. This second server has nginx installed, and there's a working WordPress blog there on a virtualhost, so I would need to setup a second site. To do this I've created the directory structure, and I've created a /etc/nginx/sites-enabled/sugar.domain.com configuration file that has the following: * server { listen 80; server_name sugar.domain.com *.domain.com; access_log /var/www/sugar/log/access.log; error_log /var/www/sugar/log/error.log info; location / { root /var/www/sugar; index index.php; } location ~ .php$ { fastcgi_split_path_info ^(.+\.php)(.*)$; fastcgi_pass backend; fastcgi_index index.php; fastcgi_param SCRIPT_FILENAME /var/www/sugar/$fastcgi_script_name; include fastcgi_params; fastcgi_param QUERY_STRING $query_string; fastcgi_param REQUEST_METHOD $request_method; fastcgi_param CONTENT_TYPE $content_type; fastcgi_param CONTENT_LENGTH $content_length; fastcgi_intercept_errors on; fastcgi_ignore_client_abort on; fastcgi_read_timeout 180; } ## Disable viewing .htaccess & .htpassword location ~ /\.ht { deny all; } } upstream backend { server 127.0.0.1:9000; } As far as I know, I need the *.domain.com parameter on the "server_name" flag, but something is crashing here: I get either a 403 Forbidden error, or I get PHP code (I can read the PHP file code in the browser, like normal text) that somehow is not executed. I've tried setting permissions to 755 inside the /var/www/sugar/ directory, and I've also set up the owner:group with a chown -R www-data:www-data /var/www/sugar/ The thing is, I don't now if my mistake is in the nginx site configuration, in my folder permissions, or in other place :( Could it be because of the main domain (www.domain.com) is hosted on other server? Do they have to be together necessarily?

    Read the article

  • Two-Hop SSH connection with two separate public keys

    - by yigit
    We have the following ssh hop setup: localhost -> hub -> server hubuser@hub accepts the public key for localuser@localhost. serveruser@server accepts the public key for hubuser@hub. So we are issuing ssh -t hubuser@hub ssh serveruser@server for connecting to server. The problem with this setup is we can not scp directly to the server. I tried creating .ssh/config file like this: Host server user serveruser port 22 hostname server ProxyCommand ssh -q hubuser@hub 'nc %h %p' But I am not able to connect (yigit is localuser): $ ssh serveruser@server -v OpenSSH_6.1p1, OpenSSL 1.0.1c 10 May 2012 debug1: Reading configuration data /home/yigit/.ssh/config debug1: /home/yigit/.ssh/config line 19: Applying options for server debug1: Reading configuration data /etc/ssh/ssh_config debug1: Executing proxy command: exec ssh -q hubuser@hub 'nc server 22' debug1: permanently_drop_suid: 1000 debug1: identity file /home/yigit/.ssh/id_rsa type 1000 debug1: identity file /home/yigit/.ssh/id_rsa-cert type -1 debug1: identity file /home/yigit/.ssh/id_dsa type -1 debug1: identity file /home/yigit/.ssh/id_dsa-cert type -1 debug1: identity file /home/yigit/.ssh/id_ecdsa type -1 debug1: identity file /home/yigit/.ssh/id_ecdsa-cert type -1 debug1: Remote protocol version 2.0, remote software version OpenSSH_5.9p1 Debian-5ubuntu1 debug1: match: OpenSSH_5.9p1 Debian-5ubuntu1 pat OpenSSH_5* debug1: Enabling compatibility mode for protocol 2.0 debug1: Local version string SSH-2.0-OpenSSH_6.1 debug1: SSH2_MSG_KEXINIT sent debug1: SSH2_MSG_KEXINIT received debug1: kex: server->client aes128-ctr hmac-md5 none debug1: kex: client->server aes128-ctr hmac-md5 none debug1: sending SSH2_MSG_KEX_ECDH_INIT debug1: expecting SSH2_MSG_KEX_ECDH_REPLY debug1: Server host key: ECDSA cb:ee:1f:78:82:1e:b4:39:c6:67:6f:4d:b4:01:f2:9f debug1: Host 'server' is known and matches the ECDSA host key. debug1: Found key in /home/yigit/.ssh/known_hosts:33 debug1: ssh_ecdsa_verify: signature correct debug1: SSH2_MSG_NEWKEYS sent debug1: expecting SSH2_MSG_NEWKEYS debug1: SSH2_MSG_NEWKEYS received debug1: Roaming not allowed by server debug1: SSH2_MSG_SERVICE_REQUEST sent debug1: SSH2_MSG_SERVICE_ACCEPT received debug1: Authentications that can continue: publickey debug1: Next authentication method: publickey debug1: Offering RSA public key: /home/yigit/.ssh/id_rsa debug1: Authentications that can continue: publickey debug1: Trying private key: /home/yigit/.ssh/id_dsa debug1: Trying private key: /home/yigit/.ssh/id_ecdsa debug1: No more authentication methods to try. Permission denied (publickey). Notice that it is trying to use the public key localuser@localhost for authenticating on server and fails since it is not the right one. Is it possible to modify the ProxyCommand so that the key for hubuser@hub is used for authenticating on server?

    Read the article

  • Nginx not working properly on subdomains

    - by javipas
    I've been trying to setup a Sugar CRM instance. I've got a domain that has its main site on a server (www.domain.com) and I've created a subdomain (sugar.domain.com), but I wnat this subdomain to be hosted on another server. This second server has nginx installed, and there's a working WordPress blog there on a virtualhost, so I would need to setup a second site. To do this I've created the directory structure, and I've created a /etc/nginx/sites-enabled/sugar.domain.com configuration file that has the following: * server { listen 80; server_name sugar.domain.com *.domain.com; access_log /var/www/sugar/log/access.log; error_log /var/www/sugar/log/error.log info; location / { root /var/www/sugar; index index.php; } location ~ .php$ { fastcgi_split_path_info ^(.+\.php)(.*)$; fastcgi_pass backend; fastcgi_index index.php; fastcgi_param SCRIPT_FILENAME /var/www/sugar/$fastcgi_script_name; include fastcgi_params; fastcgi_param QUERY_STRING $query_string; fastcgi_param REQUEST_METHOD $request_method; fastcgi_param CONTENT_TYPE $content_type; fastcgi_param CONTENT_LENGTH $content_length; fastcgi_intercept_errors on; fastcgi_ignore_client_abort on; fastcgi_read_timeout 180; } ## Disable viewing .htaccess & .htpassword location ~ /\.ht { deny all; } } upstream backend { server 127.0.0.1:9000; } As far as I know, I need the *.domain.com parameter on the "server_name" flag, but something is crashing here: I get either a 403 Forbidden error, or I get PHP code (I can read the PHP file code in the browser, like normal text) that somehow is not executed. I've tried setting permissions to 755 inside the /var/www/sugar/ directory, and I've also set up the owner:group with a chown -R www-data:www-data /var/www/sugar/ The thing is, I don't now if my mistake is in the nginx site configuration, in my folder permissions, or in other place :( Could it be because of the main domain (www.domain.com) is hosted on other server? Do they have to be together necessarily?

    Read the article

  • Apache2 Worker Starting Tons of Processes

    - by karmic
    I am installed apache2-mpm-worker and left all config files default (I've never touched them much). Is it normal that when I restart apache there is at least 20 apache processes starting? Shouldn't it be just 2 like it says in the configuration? Also, my memory seems to grow very quickly until my machine crashes. I don't have any mods installed.

    Read the article

  • MySQL server installation problems (windows)

    - by waitinforatrain
    Hi guys, I'm trying to install some CMS software (Wiccle). I was using XAMPP's MySQL but was getting a lot of errors (the same configuration works on another machine) so thought I'd install MySQL Community Edition to see if the proplem was related to the MySQL server. When I install and run the MySQL Community Edition service, however, it only works with my XAMPP password, and contains the same tables as the XAMPP install. Is there a common local database file where the database and login info is stored? Any help appreciated

    Read the article

  • Recreate/Regenerate vhost config in cpanel for one account

    - by Gabriel
    So my boss as allways was messing with cpanel on our production server (we only have production servers, which is bad enough)... so now all of the accounts is pointing to the default apache page in /usr/local/apache/htdocs instead of /home/useracc/public_html. Is there a way how to tell cpanel/WHM to recreate the vhost configuration of the account from scratch? the account has a database and emails associated with them so before i do anything i don't want to mess anything up with cpanel.

    Read the article

< Previous Page | 270 271 272 273 274 275 276 277 278 279 280 281  | Next Page >