https://bugzilla.wikimedia.org/show_bug.cgi?id=29162

--- Comment #3 from FT2 <[email protected]> 2011-05-27 17:43:31 UTC ---
To clarify, URL variants where robots.txt or header tags prohibit spidering
will probably be excluded from spidering in the first place. So Google will be
left to collate those URL variants it came across where robots,txt or header
tags _didn't_ prevent spidering -- and a "canonical" setting which states these
are all the same page. 

Ie this setting could help avoid duplicates but my guess is it probably _won't_
prevent URLs not stopped by robots.txt or header tags from being listed in
results.

-- 
Configure bugmail: https://bugzilla.wikimedia.org/userprefs.cgi?tab=email
------- You are receiving this mail because: -------
You are the assignee for the bug.
You are on the CC list for the bug.

_______________________________________________
Wikibugs-l mailing list
[email protected]
https://lists.wikimedia.org/mailman/listinfo/wikibugs-l

Reply via email to