Hi,

 When i run program "dmoz" sample given on the scrapy tutorial, not 
successfully, i got following errors, if any body knows please tell me

C:\Python27\Scripts\tutorial>scrapy crawl dmoz
2014-08-28 01:32:14+0530 [scrapy] INFO: Scrapy 0.24.4 started (bot: 
tutorial)
2014-08-28 01:32:14+0530 [scrapy] INFO: Optional features available: ssl, 
http11

2014-08-28 01:32:14+0530 [scrapy] INFO: Overridden settings: 
{'NEWSPIDER_MODULE'
: 'tutorial.spiders', 'SPIDER_MODULES': ['tutorial.spiders'], 'BOT_NAME': 
'tutor
ial'}
2014-08-28 01:32:14+0530 [scrapy] INFO: Enabled extensions: LogStats, 
TelnetCons
ole, CloseSpider, WebService, CoreStats, SpiderState
Traceback (most recent call last):
File "C:\Python27\lib\runpy.py", line 162, in _run_module_as_main
"__main__", fname, loader, pkg_name)
File "C:\Python27\lib\runpy.py", line 72, in _run_code
exec code in run_globals
File "c:\python27\Scripts\scrapy.exe\__main__.py", line 9, in <module>
File "C:\Python27\lib\site-packages\scrapy\cmdline.py", line 143, in execute
_run_print_help(parser, _run_command, cmd, args, opts)
File "C:\Python27\lib\site-packages\scrapy\cmdline.py", line 89, in 
_run_print
_help
func(*a, **kw)
File "C:\Python27\lib\site-packages\scrapy\cmdline.py", line 150, in 
_run_comm
and
cmd.run(args, opts)
File "C:\Python27\lib\site-packages\scrapy\commands\crawl.py", line 60, in 
run

self.crawler_process.start()
File "C:\Python27\lib\site-packages\scrapy\crawler.py", line 92, in start
if self.start_crawling():
File "C:\Python27\lib\site-packages\scrapy\crawler.py", line 124, in 
start_cra
wling
return self._start_crawler() is not None
File "C:\Python27\lib\site-packages\scrapy\crawler.py", line 139, in 
_start_cr
awler
crawler.configure()
File "C:\Python27\lib\site-packages\scrapy\crawler.py", line 47, in 
configure
self.engine = ExecutionEngine(self, self._spider_closed)
File "C:\Python27\lib\site-packages\scrapy\core\engine.py", line 64, in 
__init
__
self.downloader = downloader_cls(crawler)
File "C:\Python27\lib\site-packages\scrapy\core\downloader\__init__.py", 
line
73, in __init__
self.handlers = DownloadHandlers(crawler)
File 
"C:\Python27\lib\site-packages\scrapy\core\downloader\handlers\__init__.p
y", line 22, in __init__
cls = load_object(clspath)
File "C:\Python27\lib\site-packages\scrapy\utils\misc.py", line 42, in 
load_ob
ject
raise ImportError("Error loading object '%s': %s" % (path, e))
ImportError: Error loading object 
'scrapy.core.downloader.handlers.s3.S3Download
Handler': No module named win32api

C:\Python27\Scripts\tutorial>

Thanks,
Selvi Rathinam.

-- 
You received this message because you are subscribed to the Google Groups 
"scrapy-users" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to scrapy-users+unsubscr...@googlegroups.com.
To post to this group, send email to scrapy-users@googlegroups.com.
Visit this group at http://groups.google.com/group/scrapy-users.
For more options, visit https://groups.google.com/d/optout.
  • Run dmoz Deivanayaki Rathinam

Reply via email to