Author

Topic: Can we block "print pages" in robots.txt? (Read 1561 times)

newbie
Activity: 56
Merit: 0
December 06, 2012, 12:08:50 PM
#5
not really hard to remove them from the url and hit enter just saying

Humans working for computers...  Huh

theymos can make a robots.txt with these lines added:

User-agent: *
...(other stuff)
Disallow: /index.php?action=printpage
Disallow:  /index.php?wap2


Computers working for humans...  Grin
legendary
Activity: 1512
Merit: 1036
November 16, 2012, 06:42:40 AM
#4
theymos can make a robots.txt with these lines added:

User-agent: *
...(other stuff)
Disallow: /index.php?action=printpage
Disallow:  /index.php?wap2
legendary
Activity: 1358
Merit: 1003
Ron Gross
October 26, 2012, 06:19:58 AM
#3
Also, rel canonical might be useful.
donator
Activity: 2058
Merit: 1054
October 26, 2012, 03:23:28 AM
#2
Wap2 versions also appear very often.
legendary
Activity: 1358
Merit: 1003
Ron Gross
October 26, 2012, 03:19:05 AM
#1
Often, the top Google result/s for a query is a print page - https://bitcointalk.org/index.php?action=printpage;topic=106596.0

I don't know why these pages appear so high on Google (relatively), but they're irrelevant, since the same content always appears in a better form if you remove the "action=printpage" from the URL.

Can we add to the site's robots.txt file a direction instructing robots not to index anything with the string "action=printpage" in its URL?
Jump to: