Prevent bot from crawling certain areas of site.
Posted
by Skoder
on Stack Overflow
See other posts from Stack Overflow
or by Skoder
Published on 2010-02-09T00:43:32Z
Indexed on
2010/03/17
6:31 UTC
Read the original article
Hit count: 414
Hey,
I don't know much about SEO and how web spiders work, so forgive my ignorance here. I'm creating a site (using ASP.NET-MVC) which has areas that displays information retrieved from the database. The data is unique to the user, so there's no real server-side output caching going on. However, since the data can contain things the user may not wish to have displayed from search engine results, I'd like to prevent any spiders from accessing the search results page. Are there any special actions I should take to ensure that the search result directory isn't crawled? Also, would a spider even crawl a page that's dynamically generated and would any actions preventing certain directories being search mess up my search engine rankings?
edit: I should add, I'm reading up on robots.txt
protocol, but it relies on co-operation from the web crawler. However, I'd also like to prevent any data-mining users who will ignore the robots.txt file.
I appreciate any help!
© Stack Overflow or respective owner