Search Results

Search found 5738 results on 230 pages for 'seo friendly'.

Page 51/230 | < Previous Page | 47 48 49 50 51 52 53 54 55 56 57 58  | Next Page >

  • Accented characters representation in the URL

    - by Dan
    We have support for various languages in our website, including Spanish, French and Swedish. For now, the links in the site are NOT encoded before sent to the browser, sending the real accented chars (if such exists, i.e. href="www.(dot)example(dot)com/héllo.html") and not their HEX representation. This works & looks good on all browsers, including Chrome, FF and IE. However, we care great deal about SEO. We got this tip that encoding the links before sending them to the browser (so instead of linking to http://www.example.com/héllo, we will link to http://www.example.com/h%E9llo) will improve the way search engines will 'understand' the links and the keywords in the URL. This involves some work at our side, so we wanted to know if there's truth in that tip, but couldn't find anything addressing this issue.

    Read the article

  • Mix2010: Search Engine Optimization (SEO) for Microsoft Silverlight

    I had a great time today in my Mix2010 session on SEO for Silverlight.   You can find all the slides (more than I was able to cover in the talk) here.  and the the full play-by-play of the demo (include a link to the completed solution).    I started off talking why SEO matters Then I talked about three tips   You can view the site live by using Bing for my foodie Explorer Cooking Class with Joe.. or the other guys search engine equally well. ...Did you know that DotNetSlackers also publishes .net articles written by top known .net Authors? We already have over 80 articles in several categories including Silverlight. Take a look: here.

    Read the article

  • What is the best way to deal with 404s that are all trying to point to the same page that are from an external site?

    - by Lee
    I started getting 404s showing up in my Google Webmaster's Tools from a site linking to a specific category but with odd characters at the end of the url. So Something like this: http://example.com/category/puppies%EF%BC%9A.textwidget%E8%A6%81%E7%B4%A0%E7%B7%A8%E9%9B%86 Google Webmaster says that there are about 120 of these links and I can imagine there will be more to come. What is the best way to handle these links from an seo point-of-view? I have heard 301 redirecting too many links at one time can cause Google to ding the site but I don't want this site to continue posting broken links. Any help on this would be appreciated.

    Read the article

  • client website compromised, found a strange .php file. any ideas?

    - by Kevin Strong
    I do support work for a web development company and I found a suspicious file today on the website of one of our clients called "hope.php" which contained several eval(gzuncompress(base64_decode('....'))) commands (which on a site like this, usually indicates that they've been hacked). Searching for the compromised site on google, we got a bunch of results which link to hope.php with various query strings that seem to generate different groups of seo terms like so: (the second result from the top is legitimate, all the rest are not) Here is the source of "hope.php": http://pastebin.com/7Ss4NjfA And here is the decoded version I got by replacing the eval()s with echo(): http://pastebin.com/m31Ys7q5 Any ideas where this came from or what it is doing? I've of course already removed the file from the server, but I've never seen code like this so I'm rather curious as to its origin. Where could I go to find more info about something like this?

    Read the article

  • Magento Default Sitemap.xml

    - by chipShot
    Is the default magento sitemap.xml optimized as is for ecommerce products? I'm thinking about adding image links as well. Is it worth time investing in this for SEO gains? <url> <loc> http://demo.com/product.html </loc> <lastmod>2011-08-03</lastmod> <changefreq>always</changefreq> <priority>1.0</priority> </url>

    Read the article

  • Why Wikipedia doesn't appear as a referral in Google Analytics' Traffic sources?

    - by Rober
    One of my clients has a website and got not spammy backlinks in a Wikipedia article. When I test it for SEO purposes with Google Analytics (from different IPs), apparently there is no referral information. On the Real-Time view my test visit is visible but with There is no data for this view in the referrals subview. And this visits appear as (direct) / (none) on the Traffic sources view. Wikipedia is not hiding in any way its links origin, since it is shown in the server visits log. Is Google ignoring Wikipedia as a referral? Am I missing anything else? Update: Now it works, several days after the link was active. Maybe something is detecting for how long the link was there so that it doesn't work just from the beggining, as a security measure? Many visits are actually not recorded.

    Read the article

  • Aliasing resellers domain to primary domain

    - by Ashkan Mobayen Khiabani
    I have designed a website that accepts re-sellers and actually the concept of this website is having local re-sellers for each province (or should we say branches). I have designed this website in a way that anybody who has a domain, can point to our website (a record or cname). well most of the website content are the same, the only difference is that re-sellers website doesn't have some items on the main menu and may have some small descriptions of their own branch in some pages. I read that Google may ban websites with duplicate content (or which are significantly similar). I want to know will this be a problem for me? If yes, what else can I do? we have had considered asking our reseller to use iframe that loads our website but wanted that each reseller can have its own SEO and try harder but what I read about this duplicate thing worries me.

    Read the article

  • 301 redirects - can we not delete old pages?

    - by KBS
    First time here :) We have a page on the site which ranks well for an SEO term (top 5) but contains old information. We have added a new page but Google doesn't rank it that well. Information on these pages is time sensitive. Old: example.com/2013-related-information.html New: example.com/2014-related-information.html Obvious solution is to delete old page and do a 301 redirect to the new page. Now, can we still keep the old page by giving it a new URL. Step1: example.com/2013-related-information.html is redirect to example.com/2014-related-information.html Step2: example.com/2014-related-information.html is recreated with a new address such as example.com/new-2013-related-information.html What we are trying to do is to send the user to the fresh page but still not wasting the record copy if someone wants to go and dig up old page. Would appreciate help!! Cheers

    Read the article

  • Meta Refresh for change of page name and content

    - by user3507399
    Hopefully just a quick one. I've got a client that is changing the name of a workshop that they run. This means a change of url, page title for keywords that they have first page ranking on. The keywords are still relevant so what I want to avoid is a 301 redirect to a page that has different keywords to the previous page. Is the best option to keep the old page live with url and title and use a meta refresh to redirect after a period of time (not instant)? That way the SEO ranking is retained for the previous workshop name while they work on the ranking for the name change? Would a 301 redirect have an inverse effect? Thanks!

    Read the article

  • Uses of LINK tag

    - by DisgruntledGoat
    The <link> tag appears to have many uses aside from stylesheets. For example the W3 suggest using it for previous/next/index pages. I know that Opera also has a Navigation toolbar that will show links when present, including Home, Index, Contents, Previous, Next, Copyright, Author and more. (I doubt it is actually used by more than a handful of people.) Are there any other attributes that are useful, or other uses for the ones above? What about SEO benefit?

    Read the article

  • How can I create multiple mini-sites with similar/duplicate content without hurting my search engine rank?

    - by ekpyrotic
    Essential background: I run a small company that lets members of the public post handwritten letters to their local politician (UK-based). Every week a number of early stage bills (called Early Day Motions) are submitted for debate in the House of Commons, and supporters of the motion will contact their local Members of Parliament, asking them to sign the motion. The crux: I want to target these EDMs with customised mini-sites, so when people search "EDM xxx", they find my customised mini-site, specifically targeting that EDM (i.e., "Send a handwritten letter to your MP asking them to sign EDM xxx"). At the moment, all these mini-sites (and my homepage) have duplicate content with only the relevant EDM name, number, and background image changed. (For example, http://mailmymp.com and http://mailmymp.com/edm/teaching-life-saving-skills-at-school-edm-550.php). The question: Firstly, will this hurt my potential search engine ranking? And, if so, what's the best way to target these political campaigns in an efficient manner without hurting my SEO prospects?

    Read the article

  • Content Optimization only?

    - by danie7L T
    There are tons of discussions around tips&tricks to improve Search Engines "ranking" and SEOs. What if the focus of the webmaster/client is 100% set on the quality of the content with precise keywords in meta tags, clean design, regular articles updates, clean URLs and highly filtered external links leading to pages on websites dealing on the same,or related subjects; isn't it the job of a good search engine like Google to catch this website and show it in its front-page ? Or does Search Engines count on us to help them find us, and webmasters will always have to be up-to-date regarding SEO tools and rules updates on top of websites' design, browsers customization, progressive enhancement etc ?

    Read the article

  • 301 redirect to 404 page?

    - by Kristian
    Currently i'm migrating www. prefix from my urls and use htaccess to do the job. Since we have new software and cleaned database some of the old urls doesnt exists anymore. Therefore some requests redirect to 404 page. 1. www.domain.com/old-page # htaccess redirect to non-www url, 301 2. domain.com/old-page # page does not exists, 404 Does this method have any SEO issues, or even affect pagerank? Or should i check the page existence before redirecting and show 404 without redirect?

    Read the article

  • dedicated domain name VS just folders under a single domain?

    - by Ben Keating
    I run WordPress-Multisite for several sites. Each of these sites resolve under a single domain, e.g. example.com/foo/, example.com/bar/. I also have domain names for these e.g. foo.com, bar.com. which are currently redirects, so if a user hits foo.com, they are redirected (301) to example.com/foo/. My question is, should it be the other way around? should I use the dedicated domain names directly? What are the pros/cons of putting multiple sites under a single domain vs their own dedicated domains. I guess im asking with SEO and findability in mind.

    Read the article

  • problem showing my website correctly in search engines

    - by dinbrca
    Hello guys, I have a website which i have indexed on google for example (like 15 days ago). some of my pages pass arguments like: http://www.bla.com/products.php?pro=bla&page=view suddently i saw that passing arguments like this isn't good for SEO purposes and started using htaccess rewrite. and changed the arguments to like this: http://www.bla.com/products/bla/*view*/ now my site on google still shows as i showed at link number 1 what should i do? i thought i should wait for the search engine to crawl my site again but nothing happened. thanks in advanced, Din

    Read the article

  • Use of another country domain name can influence search engines results?

    - by DontVoteMeDown
    I'm studing a way to create my company domain based on it's name. Consider that my company's name is Another Store and I want to register a domain like anothersto.re - this is just an example. That domain is strictly chosen by marketing. What happens is that my company is stabilished in Brazil and our domain here is .br. The .re domain stands for an island near France so haves nothing to do with my country. If that domain is chosen what it can imply about SEO questions? Did it will have any influence on search engines results considering that they look over user's region? This kind of domain use became common between modern companies - and marketing strategies - and that is why I'm considering it.

    Read the article

  • What options do I have for a Wedding Car website? [on hold]

    - by David
    I currently run a Wedding Car website that is ranking in the top 5 for my local area, its not doing so well outside of that. I have listed the company in Google Business so it shows in the SERP and I have a Facebook and Twitter account linked, that I update regularly. I know keeping content fresh is important but nothing much changes so I am unsure what options I have? Should I create a blog and talk about the cars, weddings etc? I am totally unsure where to go with this website, the site currently has a few images and a small amount of text. Things I am considering, but would like advice on: A Mobile version of the website A Blog on the website A gallery page with pictures and descriptions Asking local companies to link to my website Essentially I have a small business website but not much content, because really there isnt! I am looking for long term organic ways to get a good seo rank.

    Read the article

  • If my URL's are static, but then parsed by Javascript, does that make it crawlable?

    - by Talasan Nicholson
    Lets say I have a link: <a href="/about/">About Us</a> But in Javascript [or jQuery] catches it and then adds the hash based off of the href attribute: $('a').click(function(e) { e.preventDefault(); // Extremely oversimplified.. window.location.hash = $(this).attr('href'); }); And then we use a hashchange event to do the general 'magic' of Ajax requests. This allows for the actual href to be seen by crawlers, but gives client-side users with JS enabled an ajax-based website. Does this 'help' the general SEO issues that come along with hashtags? I know hashbangs are 'ok', but afaik they aren't reliable?

    Read the article

  • How to recommend that Google indexes some keywords?

    - by Werewolf
    I've read many articles about SEO. I've tried to implement my knowledge on a site but I haven't gotten good results in 6 months. e.g.: I've used Google Webmaster Tools, sitemaps, title tags, keywords in paragraphs, etc. My Alexa rank is growing but Google detected some keywords that isn't my goal :-(. Is there a good way to focus on a keyword on search engines? How can I recommend Google to index some desired keywords? (They are available in my pages.)

    Read the article

  • h1 title attribute [closed]

    - by codemonkey
    Possible Duplicate: Why don't TITLE tags get indexed in google? Hello there, I have a h1 element on my page which is working great. I have also been using the title attribute on this element which I don't think has been causing much harm at all. My h1 title is "The Great Ocean Road" and the title attribute on that is "Great Ocean Road" - so it's a variation of the h1 text. These are both keywords for the site so i'm hoping it's a good thing for seo. Is that a bad idea do you think? I'm not sure what Search Engines think about using a title attribute or even if they would 'mark me down' for doing it in such a way. Do you think if the h1 text is "The Great Ocean Road", the title attribute should be "The Great Ocean Road" aswell? Thank you in advance

    Read the article

  • 404 code/header for search engines, on removed user content?

    - by mowgli
    I just got an email, from a former user on my website He was complaining that Google still shows the contact page he created on my site, even though he deleted it a month ago This is the first time in many years anyone requests this I told him, that it's almost entirely up to Google what content it wants to keep/show and for how long. If it's deleted on the site, I can't do much, other than request a re-visit from the googlebot The user-page already now says something like "Not found. The user has removed the content" TL;DR: But the question is: Should I generally add a 404 header (or other) for dynamic user content that has been removed from the site? Or could this hurt the site (SEO)?

    Read the article

  • Are people getting away with the "follow 1000s and then unfollow" Twitter trick? [closed]

    - by Baumr
    It seems that more and more people are trying to 'cheat' their way into more Twitter followers. The basic mechanism is: Follow thousands of people on Twitter with the hope that they'll follow you back. Once it reaches a point you're happy with, start gradually unfollowing them. That way, at the end of the day, it'll look like a lot of people follow you unconditionally. I've seen self-proclaimed social media and SEO experts do this. It's clear they want to look influential — and will use black hat social media tactics to do so. I can see how it can work, so is Twitter letting them get away with it? Should it?

    Read the article

  • Multi language switch links translated or in current language?

    - by FFish
    Should I do: A: translate the language links in the current language: (if I am on the English version) <a href="en/">English</a> | <a href="it/">Italian</a> | <a href="fr/">French B: the links in the native languages: <a href="en/">English</a> | <a href="it/">Italiano</a> | <a href="fr/">Français</a> From a user perspective option B is obvious, but what about SEO?

    Read the article

  • Can we 301 redirect to a new page, but still publish the old content somewhere else?

    - by KBS
    We have a page on the site which ranks well for an SEO term (top 5) but contains old information. We have added a new page but Google doesn't rank it that well. Information on these pages is time sensitive. Old: example.com/2013-related-information.html New: example.com/2014-related-information.html Obvious solution is to delete old page and do a 301 redirect to the new page. Now, can we still keep the old page by giving it a new URL. example.com/2013-related-information.html is redirected to example.com/2014-related-information.html example.com/2014-related-information.html is recreated with a new address such as example.com/new-2013-related-information.html What we are trying to do is to send the user to the fresh page but still not destroying the record copy if someone wants to go and dig up the old information.

    Read the article

< Previous Page | 47 48 49 50 51 52 53 54 55 56 57 58  | Next Page >