Search Results

Search found 5639 results on 226 pages for 'drupal seo'.

Page 168/226 | < Previous Page | 164 165 166 167 168 169 170 171 172 173 174 175  | Next Page >

  • How many pages can i add to a new website without going to the google sandbox ?

    - by François
    Hello there, I'm about to open a new website wich have 15.000 pages (ecommerce store). Of course, i will not publish all this pages at the same time, but i'm looking for infos on how much pages should i start with without going to the sandbox. For example, could i start with a 50 pages website or is it too much ? Then, have you an idea (i know there's no precises rules here) what's the frenquency / volume of pages i could add later ? (50 pages / a day is it ok ?) Thank you very much for your advice, and sorry for my english :-)

    Read the article

  • php request variables assigning $_GEt

    - by chris
    if you take a look at a previous question http://stackoverflow.com/questions/2690742/mod-rewrite-title-slugs-and-htaccess I am using the solution that Col. Shrapnel proposed- but when i assign values to $_GET in the actual file and not from a request the code doesnt work. It defaults away from the file as if the $_GET variables are not set The code I have come up with is- if(!empty($_GET['cat'])){ $_GET['target'] = "category"; if(isset($_GET['page'])){ $_GET['pageID'] = $_GET['page']; } $URL_query = "SELECT category_id FROM cats WHERE slug = '".$_GET['cat']."';"; $URL_result = mysql_query($URL_query); $URL_array = mysql_fetch_array($URL_result); $_GET['category_id'] = $URL_array['category_id']; }elseif($_GET['product']){ $_GET['target'] = "product"; $URL_query = "SELECT product_id FROM products WHERE slug = '".$_GET['product']."';"; $URL_result = mysql_query($URL_query); $URL_array = mysql_fetch_array($URL_result); print_r($URL_array); $_GET['product_id'] = $URL_array['product_id']; The original variable string that im trying to represent is /cart.php?Target=product&product_id=16142&category_id=249 And i'm trying to build the query string variables with code and including cart.php so i can use cleaner URL's So I have product/product-title-with-clean-url/ going to slug.php?product=slug Then the slug searches the db for a record with the matching slug and returns the product_id as in the code above.Then built the query string and include cart.php

    Read the article

  • how to remove .php from a certain file (with apache .htaccess)

    - by user2015253
    I want a certain file (only this!) to remove the php extension for calling it. BUT: It uses get parameters and they should remain! I tried it with my .htaccess and something like: RewriteEngine on followed by either RewriteRule ^file(.*)$ file.php$1 or RewriteRule ^file(.+)$ file.php$1 but it doesn't work. (First gives error 500, second gives 404) Example what I want to call in the browser: file?param=asd&foo=bar - It should call as: file.php?param=asd&foo=bar

    Read the article

  • What is the precision of the priority field in sitemap.xml?

    - by Christoph
    Unfortunately the specification does not tell anything about precision. The xml scheme definition states that it is of the type xsd:decimal: <xsd:restriction base="xsd:decimal"> <xsd:minInclusive value="0.0"/> <xsd:maxInclusive value="1.0"/> </xsd:restriction> I have a sitemap generator that uses up to 10 positions after decimal point. Where often only the last few positions differ. These numbers are perfectly right according to the xsd, but yet i found some pages(3,4) that state that only 0.0, 0.1, 0.2, .., 1.0 are valid values. How will the search engines react to such a sitemap? Will some just round the value? I know that it is unlikely that someone can provide an answer to that question, unless he works for that search engine, but i think experiences will also do.

    Read the article

  • seo custom domains

    - by BamBam
    I'm trying to do the following: I have website like apartments.com. sometimes, I want to expand to different cities, so for SEO purposes, I might create a separate domain like bostonapartments.com or newyorkapartments.com for only boston and new york apartments. bostonapartments.com and the main domain apartments.com are all hosted on 1 server. what I did was I used Virtual Host apache config to direct bostonapartments.com to a directory on that server, and then used an iframe to load content for bostonapartments.com from apartments.com. So all the content will be hosted on apartments.com, but bostonapartments.com will get the content from apartments.com. how can I accomplish this effectively in a scalable way using php, apache, mysql? btw, I do not own apartments.com, I'm just using that as an example.

    Read the article

  • Has Google introduced a system that allows multiple "keyword-only" domains per site?

    - by tags2k
    I've been told by a client that "a friend" told them that as of January 2010, Google allows multiple domain names that have keywords in them to be associated with a single site. To be honest this sounds rather April-foolish but I'm not sure when his "friend" told him so for the time being I have to take it at face value. I've heard nothing of this and have searched for such a thing this morning, to find nothing but warnings against this practice. Said client seems keen on buying up lots of domains today, so before he insists upon it I just want to be absolutely sure - has Google silently introduced such an allowance, or is there something else they introduced earlier this year that he could be getting confused with? Thanks for any light you can shine on this!

    Read the article

  • Is window.location.href = 'some_page.html' followed by search engines?

    - by Arkaaito
    Currently our website uses links to allow the user to change their locale. The problem with this is that you get a lot of random outlinks from each page on the site to... the same page, in other languages. When a search engine traverses this, it gets an excessively complex view of the site. We were going to change it to a form post to avoid this. However, it seems to me that we should just be able to change it to an onclick="window.location.href='change_my_language.php'" rather than an href="change_my_language.php". Am I right? Or do the major search engines scan for and follow this sort of thing nowadays?

    Read the article

  • String Functions in IIS Url Rewritting Module

    - by Nariman
    The IIS URL Rewrite Module ships with 3 built-in functions: * ToLower - returns the input string converted to lower case. * UrlEncode - returns the input string converted to URL-encoded format. This function can be used if the substitution URL in rewrite rule contains special characters (for example non-ASCII or URI-unsafe characters). * UrlDecode - decodes the URL-encoded input string. This function can be used to decode a condition input before matching it against a pattern. The functions can be invoked by using the following syntax: {function_name:any_string} The question is: can this list be extended by introducing a Replace function that's available for changing values within a rewrite rule action or condition?

    Read the article

  • checking google pagerank api

    - by Bharanikumar
    Hi , I am doing the small application for link exchange, so before i approve the URL , i want to check his/her site page rank, I know , there some site tell the Google page rank position , But am looking some API , for page rank checking... Thanks Bharanikumar

    Read the article

  • Cakephp, Route old google search results to new home page

    - by ion
    Hi there, I have created a new website for a company and I would like all the previous search engine results to be redirected. Since there were quite a few pages and most of them where using an id I would like to use something generic instead of re-routing all the old pages. My first thought was to do that: Router::connect('/*', array('controller' => 'pages', 'action' => 'display', 'home')); And put that at the very end of the routes.php file [since it is prioritized] so that all requests not validating with previous route actions would return true with this one and redirect to homepage. However this does not work. I'm pasting my routes.php file [since it is small] hoping that someone could give me a hint: Router::connect('/', array('controller' => 'pages', 'action' => 'display', 'home')); Router::connect('/company/*', array('controller' => 'articles', 'action' => 'view')); Router::connect('/contact/*', array('controller' => 'contacts', 'action' => 'view')); Router::connect('/lang/*', array('controller' => 'p28n', 'action' => 'change')); Router::connect('/eng/*', array('controller' => 'p28n', 'action' => 'shuntRequest', 'lang' => 'eng')); Router::connect('/gre/*', array('controller' => 'p28n', 'action' => 'shuntRequest', 'lang' => 'gre')); Router::parseExtensions('xml');

    Read the article

  • String Functions in IIS Url Rewrite Module

    - by Nariman
    The IIS URL Rewrite Module ships with 3 built-in functions: * ToLower - returns the input string converted to lower case. * UrlEncode - returns the input string converted to URL-encoded format. This function can be used if the substitution URL in rewrite rule contains special characters (for example non-ASCII or URI-unsafe characters). * UrlDecode - decodes the URL-encoded input string. This function can be used to decode a condition input before matching it against a pattern. The functions can be invoked by using the following syntax: {function_name:any_string} The question is: can this list be extended by introducing a Replace function that's available for changing values within a rewrite rule action or condition? Another way to frame the question: is there any way to do a global replace on a URL coming in using this module? It seems that you're limited to using regular expressions and back-references to construct strings, without a search/replace functionality to replace every X with Y in {REQUEST_URI} before issuing a redirect.

    Read the article

  • Best methods to make urls friendly?

    - by Geuis
    We're working on revising the url structure for some of our movie content, but we aren't quite sure on the best way to handle odd characters. For example, '303/302' '8 1/2 Women' 'Dude, Where's My Car?' '9-1/2 Weeks' So far, we're thinking: /movies/303-302 /movies/8-1-2-women /movies/dude-wheres-my-car /movies/9-1-2-weeks Is this the best solution? Is there anything we're forgetting?

    Read the article

  • Google says: Sort parameters in URL problematic

    - by feklee
    From Google's recommendations for URL structure: Sorting parameters. Some large shopping sites provide multiple ways to sort the same items, resulting in a much greater number of URLs. For example: http://www.example.com/results?search_type=search_videos&search_query=tpb&search_sort=relevance&search_category=25" When linking from outside, then having URLs differing only by sort parameters is obviously a bad idea: Google will not understand that these links point to the same item, i.e. that the item is popular. Therefore ranking will be lower than it should. But what's the alternative? Using a fragment identifier (#), and then doing the sorting in JavaScript? What else? Some settings in Webmaster tools?

    Read the article

  • Why is GXHC_gx_session_id appended to URLs?

    - by Nariman
    Based on my limited understanding of this parameter [1] it seems to be used in representing cookieless session IDs in java applications... but strangely, we're now noticing that a 3-year-old .NET stack is now appearing in Bing SERPs with GXHC_gx_session_id appended to the domain - and we're not alone: http://www.bing.com/search?q=GXHC_gx_session_id http://www.google.ca/#hl=en&q=GXHC_gx_session_id When comparing Google SERPs to Bing SERPs there are some inconsistencies in whether a particular site carries this parameter - is it then a bing-specific issue only? What else could cause this parameter to be appended to indexed URLs if the target environment (anything behind the load balancers) isn't running java? [1] - http://java.itags.org/java-web-tier-apis/72018/

    Read the article

  • OsCommerce - Remove osCsid in urls

    - by user203350
    I discovered that Google has indexed the osCsid parameter in some of my URLs at my osCommerce application. I tried to find a solution but I don't want to do something if I'm not sure what I'm doing. I only have access to the FTP and not the CMS system. Is it possible to solve this problem through the FTP?

    Read the article

  • Website Images not indexed by Google, Yahoo and Bing

    - by Nabil Kadimi
    Hi, My classifieds website has been present online since 2006, the html pages are indexed and rank as expected whereas a search on Google Images for site:example.com returns nothing & in Yahoo or Bing it returns only a few image results, 8 to 10. Here is an example of a response HTTP headers as reported by Firebug: Date Sat, 15 Jan 2011 20:38:21 GMT Server Apache Cache-Control max-age=34560000 Expires Sun, 19 Feb 2012 20:38:21 GMT Accept-Ranges bytes Last-Modified Fri, 14 Jan 2011 21:59:16 GMT Vary Accept-Encoding Content-Encoding gzip Content-Length 21675 Connection close Content-Type image/jpeg What should I do to tell search engines to index my website images? Thanks in advance.

    Read the article

  • How to get sites identical in content but different in language and TLD indexed by major search engi

    - by mojo77
    Hi! Is it possible to get two "editions" of a website both indexed by the major search engines (Google/Yahoo/Bing/Teoma) which differ in content language only and are hosted under different TLDs? Say English content is available at "http://domain.com/", German content at "http://domain.de/". Now, if e.g. Google.com is used I want it to list the "domain.com" entry and vice versa. Is "Duplicate Content" an issue here? Thanks so much for advising me!

    Read the article

  • how to make a website look professional in google search results

    - by altvali
    Hi all! There are a lot of websites that look professional in Google results. Try searching for 'stackoverflow' and you'll see at the top a result with a title, a description and a table of 8 links to stackoverflow categories. That's what i'm interested in producing for future websites. So what must be done? Does it depend on the number of visitors? how long does it take until the results start looking like that?

    Read the article

  • Hiding part of a page from Search Server 2010 Express

    - by Jonathan
    I'm working on a soon-to-be-public-facing site, and we want to have our search live on day 1, and want it to be searchable but non-public during testing, so we're planning to use something whose crawling we can control -- Search Server 2010 Express. However, if I search for something in my top navigation bar, I get nearly every page as a hit. It kinda makes sense, as every page has that content, but it's completely irrelevant on most pages. I want it to crawl through my navigation, but ignore the text within the navigation for search results. I was hoping that it'd just figure that out on it's own (the HTML for the top nav is static), but it's apparently not. Is there some standard thing I can put in my HTML that will achieve the effect I'm going for? On a side note: when I go live, will I have the same problem with public search engines, or do they tend to be smarter?

    Read the article

  • htaccess querystring based urls

    - by praveenmon
    hi guys, Suppose my website is www.example.com Earlier the urls of my website was like this www.example.com/php/bookdetails.php?bid=33 Last week i changed the urls of my site to this www.example.com/bookdetails/33 and wrote an htacces like this RewriteRule ^bookdetails/(.*)$ php/bookSearchResult.php?bid=$1 The problem is that there are some external site using my earlier URL structure (www.example.com/php/bookdetails.php?bid=33) and what i want is to redirect it to my new URL structure (www.example.com/bookdetails/33). I know it can be possible in htaccess, but i have no idea how to do it

    Read the article

< Previous Page | 164 165 166 167 168 169 170 171 172 173 174 175  | Next Page >