Search Results

Search found 5530 results on 222 pages for 'nested urls'.

Page 91/222 | < Previous Page | 87 88 89 90 91 92 93 94 95 96 97 98  | Next Page >

  • CommonFilter and CommonData solutions on CodePlex updated

    - by TATWORTH
    The CommonFilter and CommonData solutions on Codeplex have been updated post VS2010 SP1. The respective URLs are: http://commondata.codeplex.com/releases/view/62502 http://commonfilter.codeplex.com/releases/view/62499 CommonFilter is a cut-down version of CommonData containing just the filter functions. Common Data contains a vast number of useful functions for building ASP.NET web sites including: Lightweight reporting to a custome event log Filter functions for common types of data input

    Read the article

  • Asterix in URL?

    - by KajMagnus
    Are there any reasons I shouldn't use an asterix (*) in a URL? Background: With asterixes, I could provide these nice and user friendly (or what do you think??) URLs: example.com/some/folder/search-phrase* means search for pages with names starting with "search-phrase", located in /some/folder/. example.com/some/**/*search-phrase* means search for any page with "search-phrase" anywhere in its name. example.com/some/folder/* means list all pages in /some/folder/ (rather than showing the /some/folder/index page).

    Read the article

  • Windows 8 App Wish List

    - by Tim Murphy
    As I have been using Windows 8 more some of them apps that come with the system have been missing some features that I would like to see.  So Microsoft, here is my wish list for some new features. Skydrive Copy files from one folder to another in Skydrive Get public and read only URLs of a file   OneNote MX Print a page   Mail Accept/Reject appointments Sort inbox Search inbox (use the search charm) Print email   Store Keyword search (use the search charm)   del.icio.us Tags: Windows 8

    Read the article

  • Can a site recover by itself after dropping google page rank for 404 errors?

    - by Jeff
    Recently redid a website and changed the directory / URL structure. I did some .htaccess redirects for the main landing pages - however when reviewing web master tools received 404 errors for the rest of the changed URLs and noticed that Google dropped my site from the #1 position to around the 5th page. I corrected all the 404s by providing redirects in the .htaccess, resubmitted the site map and tested the google crawl bot. Will my page regain its rank by itself - or am I going to have to put some time into like I originally did?

    Read the article

  • Toggle CNAME entries using PHP?

    - by skibulk
    Is it possible with PHP to dynamically toggle CNAME entries? For example I have two mirrors with media to be served on my website. Mirror 1 has a monthly bandwidth cap so upon reaching it I want to automatically toggle to mirror 2. I want to use CNAME because the resulting urls appear to be identical to search engines, an SEO friendly approach. If there are SEO friendly alternatives I'd like to hear them as well.

    Read the article

  • Google Sitemap Limits?

    - by Anonymous -
    I've read in multiple places that Google's sitemap limits sits at 50,000 URLs per sitemap - though it's my understanding that you can submit multiple sitemaps to overcome this problem. I've also found that Google follow the sitemap protocol found here. My question is - is there anywhere where Google directly comment on the specifications and limits of sitemaps they accept? All the information I've found isn't behind any Google domain.

    Read the article

  • What is SEO and Why Does it Matter?

    SEO, or Search Engine Optimization, is the act of optimizing your website for search engines like Google, Bing, Yahoo and others. Back in 2008 Google's blog mentions 1 trillion unique URLs, that's incredible and according to the blog that is the number of sites they were aware of and that the actual number is infinite!

    Read the article

  • How to configure Google sitemap links in Wordpress? (without editing its HTML or PHP source code)

    - by Alexander Farber
    I run a Wordpress 3.7.1–de_DE site, but don't have much experience with it yet. When my site comes up in a Google search, there are 2 Google sitemap links displayed underneath: One of them points to a non-existent webpage /imprint though and I had to add a page at that URL to workaround this (and I want the URL actually be /impressum anyway since the site is in German and has German URLs). How to configure Google sitemap links in Wordpress (without editing its HTML or PHP source code)?

    Read the article

  • GA goal match URL regular expression

    - by MotoTribe
    I'm trying to setup a Goal URL with REGEX matching but it's not working. The Url I'm trying to match is: user/12345/edit?registration=1 with "12345" being the userid that changes. user/[0-9]*/edit?registration=1 Should work, but doesn't. When I do an advanced search for pages with RegEx match (in the new GA interface) it shows no results. If I search for "contains" /edit?registration=1 it shows all the Urls. What am I missing?

    Read the article

  • Download ASP.NET MVC 2 Release Candidate 2

    ASP.NET MVC is a free, fully supported framework that enables developers to quickly build standards-based, SEO-friendly Web sites by offering complete control over the HTML and URLs. Learn more about this release, and provide feedback to the team.

    Read the article

  • Total Solar Eclipse 13/November/2012 - update

    - by TATWORTH
    Panasonic Eclipse Live will start their broadcast at 18:30 UT tonight at https://www.facebook.com/PanasonicEclipseLive/app_435671416492320Alternative URLs are http://www.ustream.tv/channel/panasonic-eclipse-live-by-solar-power-1 and http://www.ustream.tv/channel/panasonic-eclipse-live-by-solar-power-2/collection/744e86aa753e(The start time is 04:30 by local Australian time and will be the 14/November for them.)

    Read the article

  • Does similar Titles considered duplicates relating to SEO

    - by Uri
    I have built a testing service and I wonder if I should be concerned that search engines will consider similar titles as duplicates. For example: Some URLs title can be differed from others by only one word such as "Senior" and "Junior" Title A: C++ Online Test for Juniors Title B: C++ Online Test for Seniors Another example is with the "+" sign: Title A: C Standard Library Online Test for Seniors Title B: C++ Standard Library Online Test for Seniors Should I assume search engines will understand there is a difference in the titles? And the titles are not duplicated?

    Read the article

  • Build Your Own Website For Free - Do it Yourself

    There are as many needs for a website as there are URLs on the web. You may want to put up a website to keep friends and family up to date on the happenings in your family. A business without a website on today's Internet is like a business without a purpose or a product. Here's how to do it--For Free.

    Read the article

  • Get location of "true" GVFS path from Nautilus

    - by user946850
    I mount my Samba shares through Nautilus. I know that, behind the scenes, the share is fuse-mounted and Nautilus somehow does the translation between smb:// URLs and the actual location in the file system. Is it possible to retrieve the "true" path in the file system for a GVFS-mounted share, for a directory or a file? How? This is related, but does not answer my question: How to copy the current path from Nautilus?

    Read the article

  • Which of the following is better optimized for search engines, namely, Google? [duplicate]

    - by Carnotaurus
    This question already has an answer here: How should I structure my URLs for both SEO and localization? 7 answers I'm about to configure my routes. Which of the following is better optimized for search engines, namely, Google: a) /pubs/12337/Crown_Inn/Stockport b) /pubs/12337/Crown_Inn_Stockport c) /pubs/12337/Stockport/Crown_Inn d) /pubs/Stockport/Crown_Inn.html e) /Stockport/Crown_Inn.html f) Some other variant

    Read the article

  • Hilarious web addresses revealed in new book

    <b>Telegraph:</b> "The compendium of ill thought out web addresses, largely from companies who naively slurred their innocent-sounding names into a single word without noticing the resulting double entendres, lists more than 150 "slurls", or slur URLs."

    Read the article

  • Judgment Calls in SEO Add Up to Results

    The titles and descriptions seen above the URLs on search engine results pages are taken by the search engines from the Meta data of the pages at first until other options are planted in directories during an SEO campaign. If a site has no meta description and no SEO content out there on the Web, the search engine selects some relevant snippets of content from somewhere on the site. The answer to the query may be there; if not the searcher will have to access the site and look for the information.

    Read the article

  • Nginx: Rewrite rule for subfolder

    - by gryzzly
    Hello, I have a subdomain where I want to keep projects I am working on, in order to show these projects to clients. Here is the configuraion file from /etc/nginx/sites-available/projects: server { listen 80; server_name projects.example.com; access_log /var/log/nginx/projects.example.com.access.log; error_log /var/log/nginx/projects.example.com.error.log; location / { root /var/www/projects; index index.html index.htm index.php; } location /example2.com { root /var/www/projects/example2.com; auth_basic "Stealth mode"; auth_basic_user_file /var/www/projects/example2.com/htpasswd; } location /example3.com/ { index index.php; if (-f $request_filename) { break; } if (!-f $request_filename) { rewrite ^/example3\.com/(.*)$ /example3\.com/index.php?id=$1 last; break; } } location ~ \.php { root /var/www/mprojects; fastcgi_pass 127.0.0.1:9000; fastcgi_index index.php; fastcgi_param SCRIPT_FILENAME $document_root$fastcgi_script_name; include /etc/nginx/fastcgi_params; } } I want to be able to place different php engines (wordpress, getsimple etc.) in subfolders. These engines have different querry parameters (id, q, url etc.) so in order to make preety URLs work I have to make a rewrite. However, above doesn't work. This is the response I get: Warning: Unknown: Filename cannot be empty in Unknown on line 0 Fatal error: Unknown: Failed opening required '' (include_path='.:/usr/local/lib/php') in Unknown on line 0 If I take out "location /example3.com/" rule, then everything works but with no preety URLs. Please help. The configuration is based on this post: http://stackoverflow.com/questions/2119736/cakephp-in-a-subdirectory-using-nginx-rewrite-rules I am using Ubuntu 9.10 and nginx/0.7.62 with php-fpm.

    Read the article

  • IIS6 Wildcard Mapping to ASP.NET - no file extension results in IIS 404

    - by Ian Robinson
    I'm trying to perform what I understand to be a relatively simple task. I'd like to remove the extensions from the URLs on my website. I have the proper set up in my application to handle and rewrite the URLs - the trouble is I can't get past IIS to actually get to my application without the extensions. The details: I'm running IIS6 on Windows Server 2003. I've gone into the web site for my application, gone to the home directory tab, clicked "Configuration" and added a wildcard map to the following file: c:\windows\microsoft.net\framework\v2.0.50727\aspnet_isapi.dll Which I verified is the same as what is used above in the application extensions portion by .ascx, etc. If I navigate to http://mywebsite.com/Blogs the result is as follows: HTTP/1.1 404 Not Found Content-Length: 1635 Content-Type: text/html Server: Microsoft-IIS/6.0 X-Powered-By: ASP.NET Date: Thu, 14 Jan 2010 15:04:49 GMT Which seems to be a standard IIS 404 message. If I navigate to http://mywebsite.com/Blogs.aspx I get my ASP.NET app.... How can I troubleshoot this? I feel like I've double checked everything a dozen times but to no avail. I must be missing something obvious. Update: Here are the exact instructions given by the asp.net url rewriter that I'm using: IIS 6.0 - Windows 2003 Server open property page for website / virtual directory. click the 'home directory' tab click the 'configuration' button, select the 'mappings' tab click 'insert' next to the 'Wildcard application maps' section browse to the aspnet_isapi.dll (normally at c:\windows\microsoft.net\framework\v2.0.50727\aspnet_isapi.dll) Ensure that 'check that file exists' is unchecked Click OK, OK, OK to close and apply changes Update 2: I have yet to find a resolution for this. The application does not seem to be receiving the request from IIS, any further ideas?

    Read the article

  • need help with automating a CMD java tool whcih qurries alexa AWS using batch

    - by Eli.C
    Hi everyone, I need to get all available info on 600 URLs from "Alexa Web Information Service", I downloaded the java tool and I'm able to run a single query each time with a single switch/Response Group. I would like to ask how to write a batch file that would automate the process ? the java tool runs from the CMD with the following: C:java UrlInfo (key1) (key2) (URL) (Response Group) UrlInfo - constant key1 - constant key2 -constant URL - variable (I guess I need to use the "(" sign to read from a file) Response Group - variable - (14 total, and I need to run each Response Group on each of the URLs once ) the app returns data in clear text formatted as XML after each query, here is an example: C:java UrlInfo (key1) (key2) www.url.com Rank Response: (?xml version="1.0"?) (aws:UrlInfoResponse xmlns:aws="http://alexa.amazonaws.com/doc/2005-10-05/") (aws:Response xmlns:aws="http://awis.amazonaws.com/doc/2005-07-11") (aws:OperationRequest) (aws:RequestId)ec2b6-e8ae-b392(/aws:RequestId) (/aws:OperationRequest) (aws:UrlInfoResult) (aws:Alexa) (aws:TrafficData) (aws:DataUrl type="canonical")url.com/(/aws:DataUrl) (aws:Rank)472906(/aws:Rank) (/aws:TrafficData) (/aws:Alexa) (/aws:UrlInfoResult) (aws:ResponseStatus xmlns:aws="http://alexa.amazonaws.com/doc/2005-10-05/") (aws:StatusCode)Success(/aws:StatusCode) (/aws:ResponseStatus) (/aws:Response) (/aws:UrlInfoResponse) Any help would be really appreciated Thanks and regards Eli.C

    Read the article

  • need help with automating a CMD java tool which queries alexa AWS using batch

    - by Eli.C
    Hi everyone, I need to get all available info on 600 URLs from "Alexa Web Information Service", I downloaded the java tool and I'm able to run a single query each time with a single switch/Response Group. I would like to ask how to write a batch file that would automate the process? The java tool runs from the CMD with the following: C:\>java UrlInfo (key1) (key2) (URL) (Response Group) UrlInfo - constant key1 - constant key2 -constant URL - variable (I guess I need to use the "(" sign to read from a file) Response Group - variable - (14 total, and I need to run each Response Group on each of the URLs once ) the app returns data in clear text formatted as XML after each query, here is an example: C:\>java UrlInfo (key1) (key2) www.url.com Rank Response: (?xml version="1.0"?) (aws:UrlInfoResponse xmlns:aws="http://alexa.amazonaws.com/doc/2005-10-05/") (aws:Response xmlns:aws="http://awis.amazonaws.com/doc/2005-07-11") (aws:OperationRequest) (aws:RequestId)ec2b6-e8ae-b392(/aws:RequestId) (/aws:OperationRequest) (aws:UrlInfoResult) (aws:Alexa) (aws:TrafficData) (aws:DataUrl type="canonical")url.com/(/aws:DataUrl) (aws:Rank)**472906**(/aws:Rank) (/aws:TrafficData) (/aws:Alexa) (/aws:UrlInfoResult) (aws:ResponseStatus xmlns:aws="http://alexa.amazonaws.com/doc/2005-10-05/") (aws:StatusCode)Success(/aws:StatusCode) (/aws:ResponseStatus) (/aws:Response) (/aws:UrlInfoResponse) Any help would be really appreciated Thanks and regards Eli.C

    Read the article

  • Apache, Tomcat 5 and problem with HTTP basic auth

    - by Juha Syrjälä
    I have setup a Tomcat with a webapp that uses http basic auth in some of its URLs. There is a Apache server in front of the Tomcat. I have setup Apache as a proxy like this (all traffic should go directly to tomcat): /etc/httpd/conf.d/proxy_ajp.conf: LoadModule proxy_ajp_module modules/mod_proxy_ajp.so ProxyPass / ajp://localhost:8009/ ProxyPassReverse / ajp://localhost:8009/ There is a webapp installed to root of Tomcat (ROOT.war), so I should be able to use http://localhost/ to access my webapp. But it is not working with http basic auth. The problem is that everything works until I try to access URL that are protected by the HTTP basic auth. URLs without authentication work just fine. When accessing this url via apache I am getting an error message from Apache. If I access the same URL directly from tomcat, everything works just fine. I am getting this to Apache error log: [Wed Sep 01 21:34:01 2010] [error] proxy: dialog to [::1]:8009 (localhost) failed access log looks like this: ::1 - - [01/Sep/2010:21:34:01 +0300] "GET /protected_path/ HTTP/1.0" 503 360 "-" "w3m/0.5.2" I am using: Fedora release 13 (Goddard) httpd-2.2.16-1.fc13.x86_64 tomcat5-5.5.27-7.4.fc12.noarch The basic auth is implemented in the webapp (not in Apache or Tomcat). The webapp is actually implemented in Scala/Lift, but that shouldn't matter. The auth works if I access the tomcat directly. Error message that I am getting from Apache. It is curious that the title is Unauthorized and not Internal error: Unauthorized The server is temporarily unable to service your request due to maintenance downtime or capacity problems. Please try again later. Apache/2.2.16 (Fedora) Server at my.server.name.com Port 80 It could be that Apache is seeing a some thing else than 200 OK response and thinks that it is an error when it actually should pass the received 401 Unauthorized response directly to browser. If this is the problem, how to fix it?

    Read the article

< Previous Page | 87 88 89 90 91 92 93 94 95 96 97 98  | Next Page >