Blocking in Robots.txt
Search engine bots will see the robots.txt file and when it sees to exclude a URL of the hosted domain name, this URL is no longer crawled and indexed.User-agent: Mediapartners-Google
Disallow: /search
User-agent: *
Allow: /
Sitemap: http://ravihonakamble.blogspot.com/sitemap.xml
NoIndex/Follow tag
Another way to eliminate duplicate content, is to use the Meta Robots tag noindex/follow:
Syntax: meta name="robots" content="noindex,follow"
noindex value is telling search engines not to index the page, thus eliminating duplicate content. And the follow value is telling search engines to still follow the links found on this page, thus still passing around link juice.
The 301 Redirect
Redirect 301 / http://newsite.com/blog/
The Canonical Link Tag
search engines behave in the same way how it would look at a 301 redirect. It is not going to index the duplicate content page. Only the destination page will appear in the search engine index.
Syntax:
link rel="canonical" href="http://www.demo.com/shoes/mens/nike"
All links going to the duplicate content pages will be counted as links of the main content page.
Example:
http://www.demo.com/shoes/mens/nike
http://www.demo.com/nike/shoes/mens