Search Results

Search found 4133 results on 166 pages for 'seo unfriendly'.

Page 9/166 | < Previous Page | 5 6 7 8 9 10 11 12 13 14 15 16  | Next Page >

  • SEO: Joomla Category Page Optimization + Canonical Linking

    - by Huberis
    I'm wondering how best to optimize my Joomla site's SEO. I have pages with multiple articles on each page. Either via category-type pages, or via modules. In each case, I'm not wanting users to access the articles separately from the forward facing, menu-linked pages. I understand however that Joomla still generates a url for those articles, and Google can still crawl and display these articles separate from the pages. My question is what is the best way to control this so that my users get directed only to the front-facing pages? By using the canonical element for each article to point to the front-facing page it's on? Or is there a better method? Thanks for your help!

    Read the article

  • Geolocation Blocking and SEO

    - by Mahesh
    I'm in process to block a particular geolocation from accessing my site. But i am not sure if this is going to raise any flags on my website. Basically i'm getting lots of spam from particular state and there are some webmaster copying my content without any attribution or backlink. I feel that is enough and i wish to save my bandwidth on that state or geolocation. So my question is how to block particular geolocation ? Do you have any suggestions for php scripts that help block particular geolocation ? and another question: if there are any SEO disadvantages ?

    Read the article

  • SEO penalty for landing page redirects

    - by therealsix
    Using ebay as an example- lets say I have a large number of items whose URLs' look like this: cgi.ebay.com/ebaymotors/1981-VW-Vanagon-manual-seats-seven-/250953153841 I want to give my client the ability to put links to these items on their website EASILY, without knowing or checking my URL. So I created a redirect service that will map their identifier with my URL: ebay.com/fake_redirect_service/shared_identifier9918 would redirect to the link above. This works great- my clients can easily setup these links with information they already have, and the user will see the page as usual. So on to the problem... I'm concerned that this redirecting service will have a negative impact on my SEO ranking. Having a landing page redirect you immediately to a different URL seems like something a typical spam site would do. Will this hurt me? Any better solutions?

    Read the article

  • Changing the content of a website completely, and SEO

    - by Sercan
    I have a blog running since like 10 months, which have 300 organic unique visitors daily, and now I will establish an eCommerce website on that domain. That means, I will delete all the content related to the blog. And publish new pages related to eCommerce using a different script. Content of the blog and the topic of eCommerce are also quite different. How should i do this change in terms of SEO? What should I expect in terms of Search Rankings, organic hit?

    Read the article

  • SEO Impact of using Responsive Design and the serving of different content on the same URL

    - by bmenekl
    We are currently working on redesigning our site and making it responsive. It is a search intensive site with complex functionality, a lot of search filters and a lot of content. Our mobile versions of certain pages need to hide some functionality (i.e. search filters) that exists in the desktop version and/or content (mainly blocks of text that are not necessary or are increasing page load in mobile devices). My questions is this: does the situation of having the same URL (responsive site) serving slightly different content (text and/or search filters) for certain pages in different devices affect our SEO (SERPs or otherwise)?

    Read the article

  • How does URL Rewriting affect SEO?

    - by Costa
    The following paragraph is from SEO Google Guide Google is good at crawling all types of URL structures, even if they're quite complex, but spending the time to make your URLs as simple as possible for both users and search engines can help. Some webmasters try to achieve this by rewriting their dynamic URLs to static ones; while Google is fine with this, we'd like to note that this is an advanced procedure and if done incorrectly, could cause crawling issues with your site. What makes URL re-writing implementation incorrect for GoogleBot? I am using Asp.net 3.5 framework.

    Read the article

  • SEO problem for site with 2 domains [closed]

    - by Harry
    Possible Duplicate: What is duplicate content and how can I avoid being penalized for it on my site? I have two domains pointing to the same site. I want both domains to co-exist, they share most of the same content, but they differ in design and they are aimed at different markets / rivaling communities. Is there a way to let google know that these two domains are the same site and don't cause me to get hit with a duplicate content penalty? Any other general SEO tips for this situation would also be welcomed. Thanks. Come on man, why was this closed. The linked page is completely irrelevant for me.

    Read the article

  • To change url to user friendly url

    - by German
    I'm re-factoring my asp.net application from asp.net 3.5 to 4.0. Also I'm changing url to user friendly url. Example /product.aspx?id=100 to /product-name/100 All my pages indexed by search engines and the site already 6 years online. I'm planning to do 301 redirect from old pages to new one. I want to make sure I won't loose the rank and traffic. Any suggestion how to do it properly?

    Read the article

  • Which tags to use for good SEO on the page

    - by Aaditi Sharma
    I have a event page, where it has the following items. Event Name Venue Name(s) {some cases go upto 5 or more venues} Event Info {Genre(s),Language,type(s)} Date(s) on which the event is. Event Description. Since, the Event name is unique, and present in the title, I am assigning <H1> to it. However, venue names are multiple, plus the same venue may be repeated across the page, along with dates. (Each)Event Info, is used a single time on the page Dates, are descriped in a styled manner using multiple spans, however, I am going to use a title on them. Event description is in <p> tag. So My question is which heading tags to use for a good symentic description and SEO. Also the title on the dates, which format should I keep the date in? (dd/mm/yyyy)?

    Read the article

  • Site experiencing low traffic volume between 8AM and 4PM BST

    - by BizNuge
    There may be no definitive answer to this question but I thought peer review of the problem might stimulate some ideas on the topic. We have a boutique sales site that is experiencing low volumes of traffic (both UK and international) between 8AM and 4PM BST. This seems sort of strange since our target audience for the site is UK based, and this would seem to be when people are awake and online. We are in contact with another boutique site in the same sector who don't experience this issue, so it seems kinda strange. Later on in the day we are getting traffic from the UK, as well as a fair amount of international traffic, so I'm at a loss to figure this one out. The site is fairly well optimised including:- sitemap.xml Proper caching policies across the board google merchant dublin core microdata html5 pretty urls meta and content are reviewed as an ongoing concern we have decent sitelinks for direct queries thru google on the site name a decent amount of inbound links FB, Twitter, Google +1 Google maps listing [verified] site has been selling for ~4 months and is getting ~250 users per day. So I'm not entirely sure how to explain the mid day dip in our figures.... Any ideas at all would be useful. Cheers all!

    Read the article

  • category title and affect on SEO and ranking [closed]

    - by Mark
    We are working on a jobs and skills website (similiar to Skill Pages) and are deciding on the names of categories. Rather than having loads of categories and sub-categories like, for example, Builder, Electrician, Carpenter etc, we would like to have more general and easier on the eye category names. So for example we have House, Computer, Education, Art etc. So a builder would be in category House and a few others. Will this style negatively effect our SEO and ranking? And if so, should we abandon and go back to traditional categories and sub-categories?

    Read the article

  • SEO regarding using multiple <h1> tag?

    - by user622378
    Is is true <h1> tag should not have more than 1 on every page for SEO? Every pages on the header, it includes and <h1> tag. Website Name which show an image from logo class, like this: <div id="header"> <h1 class="logo"> <a href="#">Website Name</a> </h1> </div> and I also include <h1> on the contact page, help, etc and article page for example: <h1>Contact Us</h1> <h1>Name of the Article Title here</h1> On the homepage, it just has one <h1> for the logo Site Name... other pages have 2 <h1> tags

    Read the article

  • SEO - Hidden content before main site content

    - by 0pt1m1z3
    I have a two hidden divs before my main site content, one with the login form and another with the signup form. I then have login and signup buttons within the page that use JQuery to show or hide these divs. I like the effect this setup offers, dropping down from the top of the page and pushing the rest of the content down. However, recently I have been getting serious about SEO and I am wondering if these divs have been affecting my SERP rankings. Basically, every non-logged page (everything bots see) has the same two display:none; divs at the top of the document flow. Is it bad? Should I re-engineer these forms and the way they are displayed?

    Read the article

  • How to maximize SEO of an internationalized website?

    - by alditis
    I am developing a website that should be multi languages. I would like your opinions about what the best way to do it, considering get the most SEO possible. I have these alternatives: Altarnative 1: Separate domains http://www.miweb.com -- English for default http://www.miweb.com.fr http://www.miweb.com.es http://www.miweb.com.it Altarnative 2: Sub domains http://www.miweb.com -- English for default http://fr.miweb.com http://es.miweb.com http://it.miweb.com Altarnative 3: Sub folders http://www.miweb.com -- English for default http://www.miweb.com/fr http://www.miweb.com/es http://www.miweb.com/it I would like to know about your experiences. I hope the question is okay. Any suggestions, comments, objections or idea is welcome. Thank you.

    Read the article

  • category title and affect on SEO and ranking

    - by Mark
    We are working on a jobs and skills website (similiar to Skill Pages) and are deciding on categories. Rather than having load of categories like, for example, Builder, Electrician, Carpenter etc, we would like to have one word more general categories. So for example we have House, Computer, Education, Art etc. So a builder would be in category Home and a few others. Will this style negatively effect our SEO and ranking? And if so, should we abandon and go back to traditional categories and sub-categories?

    Read the article

  • seo for complex url

    - by Mark
    I have a jobs website. There are categories and sub categories e.g the url for builders is www.mydomain.com/jobs/all/active/202/tradesmen/799/builders/county/all Because of the framework im using (Elgg), the "all" and "active" come after the plugin name. Also, i need to do a lot of things with the url: get the pplugin (job), get the correct type of entities (all and active), get the category (202 and tradesmen), get the subcategory (799 & builders) and get the location (county and all). Is this URL seo friendly or is it too messy? Is something like www.mydomain.com/jobs/tradesmen/builders/all better?

    Read the article

  • Removing "www." from domain name and SEO

    - by TecMan
    We are doing a big redesign work on our website, and at least 50% of the website folders will be moved to new places with different names (i.e. many URL will be changed). Sure, Google will need some time to index new pages and we expect our SERP positions will be not so good as they are now for some time. We also have an old idea to remove www from our domain name. It seems, it's the right time to do these two works together with publishing the website with updated contents. Or is it better from SEO perspective first publish the new contents, and only after some time, when our SERP positions will return to prior results, tell Google that the domain name without www is our preferred domain name?

    Read the article

  • How to manually list set of urls for search engines to index

    - by MarutiB
    So I have created a video website which has thousands of videos and thousands of videos get added to it on a daily basis. Here is my problem :- I have created a website which basically loads the skeleton in html and puts all the content through javascript and Ajax. The problem is search engines aren't going anywhere except for the home page. Is there a way say in robots.txt where i give a link to a single html which has links to all these videos ? I agree my site is not accessible for a non-javascript user but stats show that this ratio is very low ( 0.2 % ). Is there a way I can still keep the complete AJAX website and still get each individual videos listed on google ?

    Read the article

  • Is this the place to ask SEO questions [on hold]

    - by user39206
    I'm adding business listings to an existing website of which Google has sort of been referencing well but my Meta tags have a lot to be desired! Due to other work commitments I have kind of neglected the site but now think I should put the proper time and effort into getting it just right as it does earn me money. I am a developer so I have all the necessary skills to build the site, it's almost done!. Now I'm just a little worried that I could do something wrong and loose the rapport I've built up with Google in the past or be ban. Especially with the new tags I see like: <link rel=”publisher” href=”https://plus.google.com/[YOUR BUSINESS G+ PROFILE HERE]“/> Really I just want to know for now, is this the place to ask SEO question?

    Read the article

  • How to post SEO jobs in the right way and other questions [closed]

    - by mdivk
    I have websites that needs an SEO expert to make them on the top of Google searching result based on some keywords, or at least the first page, the websites are created using common open source platforms (wp, joomla, or magento, zencart) I want to post a job on oDesk but I have no idea on how this particular type of requirement works? what would the guy do on my website? will I be asked to spent extra advertisement money? How do I know s/he really did anything on my site and what have been done? if it's just to optimize URL to make it meaningful then I don't need that, and I am sure giving every page's URL an alias is not enough to make the site popular. Beside, I want to know how much it roughly costs and how long it takes. It would be great if you can shed me some light, thank you in advance.

    Read the article

  • How do I make the home page of the website to come up in the rankings than the internal pages? [closed]

    - by Shahab
    Possible Duplicate: What are the best ways to increase your site's position in Google? Suppose I have a website, e.g. www.example.com that comes at number 6 on the Google search rankings. But the internal pages of the website i.e. www.example.com/index.php?a=1&b=2 or www.example.com/index.php comes at number 2 of the rankings. How would I make my prime domain name www.example.com to come at the top of the list ? Any guidance would be appreciated.

    Read the article

  • Pagination, Duplicate Content, and SEO

    - by Iamtotallylost
    Please consider a list of items (forum comments, articles, shoes, doesn't matter) which are spread over multiple pages. Different sort orders are supported (by date, by popularity, by price, etc). So, an URL might look like this (I use the query style here to simplify things): /items?id=1234&page=42&sort=popularity /items?id=1234&page=5&sort=date Now, in terms of SEO, I think I should be worried about duplicate content. After all, each item appears at least as many times as there are sort orders. I've seen Matt Cutts talking about the rel=canonical link tag, but he also said that the canonical page should have very similar content. But this is not the case here because page #1 in a non-canonical sort order might have completely different items than page #1 in the canonical sort order. For a given non-canonical page, there is no clear canonical page listing all the same items, so I think rel=canonical won't help here. Then I thought about using the noindex meta tag on all pages with non-canonical sort order, and not using it on all pages with canonical sort order. However, if I use that method, what will happen with backlinks that are going to non-canonical pages -- will they still spread their page rank juice, even though the first page googlebot (or any other crawler) is going to encounter is marked as "noindex"? Can you please comment on my problem and what you think is the best solution? If you think you have a better solution, please consider that 1) I do not want to use Javascript for this, 2) I do not want all the items to be on one page. Thank you.

    Read the article

  • SEO - Google and link cleaning / cloaking [closed]

    - by Jens Törnell
    Possible Duplicate: Does the Google spider render JavaScript? This a SEO related question, not a code related one. Googles own link cleaning / cloaking Gå to http://www.google.com and search for something. Hover the title and you will se a link to the page you want to go to. The URL you see when hovering is NOT the link you are clicking on. Instead of clicking you can drag the title a little bit and then hover it. Then you will se the real URL. My own link cleaning / cloaking Go to http://jsfiddle.net/NvmER/1/ and click the link, or look at the code below. You will be "redirected" to http://www.test.com. The real link are http://www.test.com/?event=23 Working code in case jsfiddle don't work If you need to se how it works I pasted a code below. <a class="direct" href="http://www.test.com/?event=23" data-redirect="http://www.test.com">Länk</a>? $(document).ready(function() { $("a.direct").live("mousedown", function(e){ var oldurl = $(this).attr('href'); var newurl = $(this).attr('data-redirect'); $(this).attr('href', newurl); }); });? Question Is this ok with Google? It's done with javascript. If you have an answer, link to a source or test to support it.

    Read the article

  • ASP.NET MVC Controllers & Actions In Regards To URLs And SEO

    - by user1066133
    The general idea is that if I were to create an MVC site, simple pages such as the contact and about pages will be placed under the Home Controller. So my URL would look like http://www.mysite.com/home/contact, and http://www.mysite.com/home/about. The above works just fine, but I really don't like the idea of having the "home" portion in the URL. So what negatives would there be if I decided to make a controller name of Contact and About and just added a single Index action so that way the URL would be simplified to http://www.mysite.com/contact and http://www.mysite.com/about. This method looks cleaner. Do any of you do the same or something similar? I've been trying to work on SEO for an escort service site I've developed and when you search for the females the link looks like http://www.mysite.com/escorts/female-escorts, and like-wise for males. I'm wondering if I should remove the Escorts Controller and just create a Female_Escorts Controller with an Index Action only so it comes out like the above as http://www.mysite.com/female-escorts.

    Read the article

  • Handling SEO for Infinite pages that cause external slow API calls

    - by Noam
    I have an 'infinite' amount of pages in my site which rely on an external API. Generating each page takes time (1 minute). Links in the site point to such pages, and when a users clicks them they are generated and he waits. Considering I cannot pre-create them all, I am trying to figure out the best SEO approach to handle these pages. Options: Create really simple pages for the web spiders and only real users will fetch the data and generate the page. A little bit 'afraid' google will see this as low quality content, which might also feel duplicated. Put them under a directory in my site (e.g. /non-generated/) and put a disallow in robots.txt. Problem here is I don't want users to have to deal with a different URL when wanting to share this page or make sense of it. Thought about maybe redirecting real users from this URL back to the regular hierarchy and that way 'fooling' google not to get to them. Again not sure he will like me for that. Letting him crawl these pages. Main problem is I can't control to rate of the API calls and also my site seems slower than it should from a spider's perspective (if he only crawled the generated pages, he'd think it's much faster). Which approach would you suggest?

    Read the article

< Previous Page | 5 6 7 8 9 10 11 12 13 14 15 16  | Next Page >