Lawrence D'Oliveiro wrote:
> In message <[EMAIL PROTECTED]>, John Nagle
> wrote:
>
>> For some reason, Python's parser for "robots.txt" files
>> doesn't like Wikipedia's "robots.txt" file:
>>
>> >>> import robotparser
>> >>> url = 'http://wikipedia.org/robots.txt'
>> >>> chk = robotparser.RobotFileParser()
>> >>> chk.set_url(url)
>> >>> chk.read()
>> >>> testurl = 'http://wikipedia.org'
>> >>> chk.can_fetch('Mozilla', testurl)
>> False
>> >>>
>
> >>> chk.errcode
> 403
>
> Significant?
>
Helpful. Also an undocumented feature. See
http://docs.python.org/lib/module-robotparser.html
John Nagle
--
http://mail.python.org/mailman/listinfo/python-list