Search Results

Search found 8370 results on 335 pages for 'seo friendly urls'.

Page 242/335 | < Previous Page | 238 239 240 241 242 243 244 245 246 247 248 249  | Next Page >

  • Web Server Scripting Hack to Maintain State and Keep a Domain Cookieless

    - by jasonspalace
    Hello, I am looking for a solution on a LAMP server to keep a site cookieless such as "example.com", where static content is served from "static.example.com", and with rules in place to rewrite requests for "www.example.com" to "example.com". I am really hoping to avoid setting up a cookieless domain for the static content due to an unanswered SEO concern with regards to CNAMEing to a CDN. Is there a way, (or safe hack), that can be implemented where a second domain such as "www.example2.com" is CNAMEd, aliased, or otherwise used with "example.com" to somehow trick a php application into maintaining state with a cookie dropped on "www.example2.com" therefore keeping all of "example.com" cookieless? If such a solution is feasible, what implications would exists with regards to SSL and cross-browser compatibility other than requiring users to accept cookies from 3rd party domains and possibly needing an additional SSL to keep the cookie secure? Thanks in advance to all.

    Read the article

  • Website Content via remote mysql or API for clients

    - by Kannan
    I would like to know whether there will be problem for shared hosting/vps/dedicated users if they request content via remote mysql or via http API. ? What will be issues if they make every request to production server for delivery of the website articles along with image urls etc. Will it cause disk wait on shared/vps/dedicated hosting users ? Will both remotemysql or http api or curl cause high CPU usage, I/O wait or Memory usage ? What should be best for speed remote mysql or http web api ? or any alternatives ? Thanku Kannan

    Read the article

  • How do I configure multiple domain names on my IIS server?

    - by Dillie-O
    We have a few websites that we are running on one instance of IIS that need to be mapped for each of their domain names. For example. Site A has the domain name coolness.com Site B has the domain name 6to8Weeks.com Site C has the domain name PhatTech.com When I look at the "Web Site Identification" section of the IIS configuration window, I notice that I can specify an IP address and port, but if I click the Advanced button, I can also configure the site based on host header values as well. How do I configure each site in IIS? Ideally I would like them to all be able to listen to port 80, so I don't have weird URLs, but I'm not sure if I do this using headers, IP addresses, both, or something else.

    Read the article

  • Git version control with multiple users

    - by ignatius
    Hello, i am a little bit lost with this issue, let me explain you my problem: I want to setup a git repository, three of four users will contribute, so they need to download the code and shall be able to upload their changes to the server or update their branch with the latest modifications. So, i setup a linux machine, install git, setup the repository, then add the users in order to enable the acces throught ssh. Now my question is, What's next?, the git documentation is a little bit confusing, i.e. when i try from a dummy user account to clone the repository i got: xxx@xxx-desktop:~/Documentos/git/test$ git clone -v ssh://[email protected]/pub.git Initialized empty Git repository in /home/xxx/Documentos/git/test/pub/.git/ [email protected]'s password: fatal: '/pub.git' does not appear to be a git repository fatal: The remote end hung up unexpectedly is that a problem of privileges? need any special configuration? i want to avoid using git-daemon or gitosis, sorry, maybe my question sound silly but git is powerfull but i admit not so user friendly. Thanks Br

    Read the article

  • Is Flash typically slow on Linux?

    - by CSarnia
    Specifically, I'm running Mint 8 (Helena). I'm extremely new to Linux, and was searching for a solution that was user-friendly and GUI oriented. The box won't be used for much other than web browsing and word processing. Anyway, it runs relatively smoothly, except for Youtube videos... especially full-screen, which runs at like 1 FPS, and even after closing, slows Firefox to a crawl until I restart it. I'd seen an xkcd comic on the matter, but regarded it as a joke until now. Is this actually a problem? Are there any remedies I can try to smooth the applications?

    Read the article

  • Can I config different VIP per service pointing same Real servers with Cisco ACE?

    - by Kamome
    I'm using Cisco ACE Module with Cisco Catalyst 6504. There are 2 real servers connected to Switch, and runs three different services. (TCP 82, 83 and 84) I have to use three different URLs per service. (for example, a.com for TCP 82, b.com for TCP 83 and c.com for TCP 84) Therefore, I need three IPs. The problem is, is it possible that configure 3 different VIPs with same real servers with Cisco ACE Module? I mean, if VIPs are 1.1.1.1, 2.2.2.2, 3.3.3.3 and real servers are 4.4.4.4, 5.5.5.5, can I configure as following : VIP 1.1.1.1:82 Real Server 4.4.4.4:82, 5.5.5.5:82 VIP 2.2.2.2:83 Real Server 4.4.4.4:83, 5.5.5.5:83 VIP 3.3.3.3:84 Real Server 4.4.4.4:84, 5.5.5.5:84 I think, logically, it's possible. But is it REALLy possible?

    Read the article

  • nginx 500 error instead of 404

    - by arby
    I have the following nginx configuration (at /etc/nginx/sites-available/default) server { listen 80; ## listen for ipv4; this line is default and implied listen [::]:80 default ipv6only=on; ## listen for ipv6 root /usr/share/nginx/www; index index.php index.html index.htm; server_name _; location / { try_files $uri $uri/ /index.html; } error_page 404 /404.html; location ~ \.php$ { try_files $uri =404; fastcgi_split_path_info ^(.+\.php)(/.+)$; fastcgi_pass unix:/tmp/php5-fpm.sock; fastcgi_index index.php; include fastcgi_params; } location ~ /\.ht { deny all; } } Instead of a 404 error, I'm getting 500 server errors on broken urls. How can I correct this?

    Read the article

  • What response should be made to a continued web-app crack attempt?

    - by Tchalvak
    I've issues with a continuous, concerted cracking attempt on a website (coded in php). The main problem is sql-injection attempts, running on a Debian server. A secondary effect of the problem is being spidered or repeatedly spammed with urls that, though a security hole has been closed, are still obviously related attempts to crack the site, and continue to add load to the site, and thus should be blocked. So what measures can I take to: A: Block known intruders/known attack machines (notably making themselves anonymous via botnet or relaying servers) to prevent their repeated, continuous, timed access from affecting the load of the site, and B: report & respond to the attack (I'm aware that the reporting to law enforcement is almost certainly futile, as may be reporting to the ip/machine where the attacks are originating, but other responses to take would be welcome).

    Read the article

  • NSD Daemon (DNS SERVER)

    - by Catalin
    While searching for a secure dns server I came across this http://www.nlnetlabs.nl/projects/nsd/ I was really impressed by what seemed to me the best option out there that's open source. One problem thought their 'tutorial' is really NOT noob friendly I have basic DNS knoledge but what's in there is out of my league. Any help in setting up this DNS server please ? I need to have multiple sites on this CentOs server I've recently got my hands on :). They also need to receive email. Details: I have a master host and would love to set this in the way described in the rows that follow: masterhost.com -> ns1.masterhost.com mail.masterhost.com www.masterhost.com addonhost.com -> ns1.masterhost.com mail.masterhost.com www.addonhost.com And so on, all answers and suggestions are welcomed, Thank you in advance.

    Read the article

  • Laptop wakes from sleep, once, due to audio controller (Windows 7)

    - by stijn
    The laptop is a recent Dell XPS 15z and the problem is as follows (reproducible about 90% of tries): put laptop to sleep using either Start-Sleep or closing the lid laptop goes to sleep after about 5 seconds, but instantly wakes again showing a black screen (touching the keyboard or moving the mouse shows the login screen one normally gets after wake) login again, put laptop to sleep latop stays in sleep mode output of powercfg -lastwake after the first instant wake shows the audio controller is responsible. Why would that be, why only the first try, and how to fix this? Wake History Count - 1 Wake History [0] Wake Source Count - 1 Wake Source [0] Type: Device Instance Path: PCI\VEN_8086&DEV_1C20&SUBSYS_04461028&REV_05\3&11583659&0&D8 Friendly Name: Description: High Definition Audio Controller Manufacturer: Microsoft

    Read the article

  • Standalone firewall + antivirus or combined security tools?

    - by pukipuki
    For years I'm using some antivirus software and different firewall. Now every antiviruses have got some firewall features and there are complete "internet security" complexes... and every firewall get some antivirus functionality and there are "internet security" versions. Firstly, it is hard and sometimes impossible to install and use standalone AV and FW. Sometimes I can't avoid them (i can't install KAV2010 without removing Outpost firewall etc). Secondly, complex solutions have some disbalance. Farewall from famous antivirus-brand is so user-friendly that is not suitable for me (lack of details in Norton Internet Security for example) and antiviruses from famous firewall-brands are still weak, it is proved by tests. What is today best-practices in case of functionality and security?) Some internet-security complex or two standalone applications from different vendors?

    Read the article

  • nginx regex characters that require quoting?

    - by Michael Louis Thaler
    So I was configuring nginx today and I hit a weird problem. I was trying to match a location like this: location ~ ^/([0-9]+)/(.*) { # do proxy redirects } ...for URLs like "http://my.domain.com/0001/index.html". This rule was never matching, despite the fact that it by all rights should. It took me awhile to figure out, based on this documentation, that some characters in regexes need to be quoted. The problem is, the documentation is for rewrites, and it specifically calls out curly braces, not square brackets. After a fair bit of experimentation that involved a lot of swearing, I discovered that I could fix the problem by quoting the regex like so: location ~ "^/([0-9]+)/(.*)" { # do proxy redirects } Is there a list somewhere of characters that nginx requires quoting regexes with? Or could there be something else going on here that I'm totally missing? This is my first nginx configuration job, so it's very possible I've misunderstood something...

    Read the article

  • Nginx trailing slash not working

    - by user1573604
    I am using the below to match urls and add a trailing slash. Under location / rewrite ^([^.]*[^/])$ $1/ permanent; According to he logs below, it is working, however the get request is still he old url, prior to adding the slash, which results in a 404. 2013/11/08 12:33:19 [notice] 29919#0: 92 "^([^.][^/])$" matches "/foo", client: 100.100.100.100, server: _, request: "GET /foo HTTP/1.1", host: "100.100.100.100" 2013/11/08 12:33:19 [notice] 29919#0: *92 rewritten redirect: "/foo/", client: 100.100.100.100, server: _, request: "GET /foo HTTP/1.1", host: "100.100.100.100" Any suggestions as to why this might be happening?

    Read the article

  • Computer and internet use on a boat?

    - by dlamblin
    Let's assume that I seriously would like to be able to use my computer as easily and as carefreely while on a boat as I do while at home. In short, if I were on a boat for extended periods, such as two weeks at a time and occasionally over a few days out of port (this would be a 2 person sailing vessel), what would my options be for power, durability, and of course, internet access? I want more than just email, as I will likely keep doing a tad of development, and will be looking at google earth from time to time. I am already assuming that we're going with some kind of laptop, maybe even a macbook pro 12". I personally feel that netbooks are underpowered. And the last computer that I got a boat-friendly rating for was a ZX Spectrum, which was before the internet, and was likely due to it's 12vdc-happiness.

    Read the article

  • How to track url redirects in browser?

    - by Prashant
    I have typed http://example.com/load/ in my browser window and pressed "ENTER" key. Now on press of enter this website redirects me to http://example.com/load/1/ and then http://example.com/load/2/ and then I finally landed on this url http://example.com/load/3/. These redirection happens at website end, I am not aware where I am going. But I finally landed on this url: http://example.com/load/3/ I want to track where all my browser gone all urls, I am not seeing it in my history as its redirect at website end. Is there any firefox addon or some toll which can track this for me? I am not sure where to ask this question, so asking it here, moderators please check!

    Read the article

  • Varnish Error 503 Service Unavailable

    - by alfish
    On Debian I have Varnish cache in front of apache and get 503 error when I try to get urls which are rarely fetched. I've put 10 min timeout in default.vlc: backend default { .host = "127.0.0.1"; .port = "81"; .connect_timeout = 600s; .first_byte_timeout = 600s; .between_bytes_timeout = 600s; } Also I have in /etc/default/varnish DAEMON_OPTS="-a :80 \ -T localhost:6082 \ -b localhost:81 \ -u varnish -g varnish \ -S /etc/varnish/secret \ -p thread_pools=7 \ -p thread_pool_min=200 \ -p thread_pool_max=5000 \ -p listen_depth=2048 \ -p thread_pool_add_delay=2\ -p lru_interval=1800 \ -h classic,169313 \ -p session_linger=100 \ -p sess_workspace=262144 \ -p connect_timeout=600 \ -p max_restarts=6 \ -s malloc,7G" I appreciate your hints to resove this problem.

    Read the article

  • Why does httpd handle requests for wrong hostnames in SSL mode?

    - by Manuel
    I have an SSL-enabled virtual host for my sites at example.com:10443 Listen 10443 <VirtualHost _default_:10443> ServerName example.com:10443 ServerAdmin [email protected] ErrorLog "/var/log/httpd/error_log" TransferLog "/var/log/httpd/access_log" SSLEngine on SSLProtocol all -SSLv2 SSLCipherSuite HIGH:MEDIUM:!aNULL:!MD5 SSLCertificateFile "/etc/ssl/private/example.com.crt" SSLCertificateKeyFile "/etc/ssl/private/example.com.key" SSLCertificateChainFile "/etc/ssl/private/sub.class1.server.ca.pem" SSLCACertificateFile "/etc/ssl/private/StartCom.pem" </VirtualHost> Browsing to https://example.com:10443/ works as expected. However, also browsing to https://subdomain.example.com:10443/ (with DNS set) shows me the same pages (after SSL certificate warning). I would have expected the directive ServerName example.com:10443 to reject all connection attempts to other server names. How can I tell the virtual host not to serve requests for URLs other than the top-level one?

    Read the article

  • How do I objectively measure an application's load on a server

    - by Joe
    All, I'm not even sure where to begin looking for resources to answer my question, and I realize that speculation about this kind of thing is highly subjective. I need help determining what class of server I should purchase to host a MS Silverlight application with a MSSQL server back-end on a Windows Server 2008 platform. It's an interactive program, so I can't simply generate a list of URLs to test against, and run it with 1000 simultaneous users. What tools are out there to help me determine what kind of load the application will put on a server at varying levels of concurrent users? Would you all suggest separating the SQL server form the web server, to better differentiate the generated load on the different parts of the stack?

    Read the article

  • Employee Monitoring software

    - by nute
    I am looking for an employee monitoring solution, that would allow us to remotely connect to our computers to see what is happening live, and preferably having some recording capabilities such as snapshots, URLs visited, etc ... I've looked around the web and most softwares I found were from unknown companies, had crappy websites, and made me feel like their either wanted me to install a virus on my computer, or to scam me. Most also seemed to have planted "reviews" online most likely written by themselves. Basically, anyone has experience with a trustworthy company to accomplish that? Thanks

    Read the article

  • Make Chrome's Omnibar behave more like the Firefox AwesomeBar

    - by Agnel Kurian
    One of my favorite features of the Firefox AwesomeBar is that I can simply type a substring of any URL or page title in my history and it finds all matches sorted by how frequently they were accessed. Example: I simply type "ask" when I want to ask something on stackoverflow.com., "inbox" goes to my GMail Inbox and so on because the substring matches any part of the URL or the page title. Chrome's Omnibar is quite frustrating in this area. I am not able to predict what it's gonna fetch and I seem to have no way to train the thing to do my bidding. I have unchecked the option that says: "Use a suggestion service to help complete searches and URLs typed..." but there has been no noticeable improvement. Any clues how I can make the Omnibar behave?

    Read the article

  • Why would image thumbnails not be showing in all search engines in all browsers?

    - by Edward Tanguay
    For over a week now, when I search for a word at Google, it correctly gives me some preview image thumbnails: But when I click on "images", it doesn't show me any thumbnails: The same thing happens at Bing.com, when I search Bing itself it gives me some thumbnails on the general search result page: But when I click on "images", it doesn't show me any thumbnails: The same thing happens at Yahoo: If I click on one of the broken thumbnails, it shows me the picture fine: Thumbnails at youtube also work fine: It seems each search engine is hanging on different URLs as shown above: t1.gstatic.com ts3.mm.bing.net thm-a02.yimg.com so it doesn't seem to be the problem of one specific URL not sending thumbnails, it is just a problem with search engine image thumbnails in general. Also, this happens in every browser I try: Explorer, Firefox, Chrome. What could be the problem? Is it my computer, some setting somewhere, my router, my Internet provider (T-Online, Germany)? Has anyone ever had this problem and solved it?

    Read the article

  • Using FTP to update files on a server

    - by Neville
    I know the FTP username and password for a site we own and need to know how we can update some files on the server. It seems quite a small thing to do and I'd like to have a go at doing it myself. A few years ago a friendly local guy help set up a website for my wife's floristry business. The site has a "contact us" page, and messages are forwarded to our home email address. We've now just changed our home email, and so I now need to reset the forwarding function on the website. The helpful local guy seems to have moved away, or retired - there's no way I can find him now. I tried to get help on how to change the forwarding address from the hosting people, but they say they can't help me. How do I go about updating the pages on the site? A step-by-step guide on how to do it would be great.

    Read the article

  • rewrite rule does not rewrite url as expected

    - by user1708687
    I have a problem with a CMS website, that normally generates readable urls. Sometimes it happens that navigation links are shown as www.domain.com/22, which results in an error, instead of www.domain.com/contact. I have not found a solution for this yet, but the page is working if the url is www.domain.com/index.php?id=22. Therefore, I'm trying to rewrite www.domain.com/22 to www.domain.com/index.php?id=22 and I have used this rewrite rule: RewriteRule ^([1-9][0-9]*)$ index.php?id=$1 [NC] I tested it using http://htaccess.madewithlove.be and here it shows the correct result, but on the website no rewrite is happening.

    Read the article

  • Nginx static files exclude one or some file extensions

    - by Evgeniy
    I'm serving up a static site via nginx. location ~* \.(avi|bin|bmp|dmg|doc|docx|dpkg|exe|flv|gif|htm|html|ico|ics|img|jpeg|jpg|m2a|m2v|mov|mp3|mp4|mpeg|mpg|msi|pdf|pkg|png|ppt|pptx|ps|rar|rss|rtf|swf|tif|tiff|txt|wmv|xhtml|xls|xml|zip)$ { root /var/www/html1; access_log off; expires 1d; } And my goal is to exclude requests like http://connect1.webinar.ru/converter/task/. Full view is like http://mydomain.tld/converter/task/setComplete/fid/34330/fn/7c2cfed32ec2eef6788e728fa46f7a80.ppt.swf. Despite the fact these URLs ends in such a format they are not static, but fake script requests, so I have a problems with them. What is the best way to do this? How can I add an exclusion for this URL or maybe I can to exclude the specific file exptension (.ppt.swf, pptx.swf) from the list of this Nginx location? Thanks.

    Read the article

  • How can I configure firefox to open links in the same window, but requests from external application

    - by Mnementh
    I hate it, when sites decide for me, which links should open in a new window, and which in the same. The back-button doesn't work. Good thing is, firefox has the option browser.link.open_newwindow. If I set this to 1, all links with target=blank open in the same window, as it should be. But now also clicks in external programs (like the email-client or newsreader) on links open this in the same window, destroying the already opened website. How can I configure firefox to open links in a website always in the same window, but in external programs opened URLs always in a new one?

    Read the article

< Previous Page | 238 239 240 241 242 243 244 245 246 247 248 249  | Next Page >