Hi, can anyone help me with my robot.txt. My contents for the page reads as
follows:

User-agent: *
Disallow: /Help
Disallow: /MediaWiki
Disallow: /Template
Disallow: /skins/

But it is blocking pages like:

   - http://www.dummipedia.org/Special:Protectedpages
   - http://dummipedia.org/Special:Allpages


and external pages like:

   - http://www.stumbleupon.com/
   - http://www.searchtheweb.com/

As you can see, my robot.txt did not block these pages. Also, should I block
the print version to prevent what Google calls "duplicate content"? If so,
how?

Response will be very much appreciated.

PM Poon
_______________________________________________
MediaWiki-l mailing list
[email protected]
https://lists.wikimedia.org/mailman/listinfo/mediawiki-l

Reply via email to