Hi, When i run program "dmoz" sample given on the scrapy tutorial, not successfully, i got following errors, if any body knows please tell me
C:\Python27\Scripts\tutorial>scrapy crawl dmoz 2014-08-28 01:32:14+0530 [scrapy] INFO: Scrapy 0.24.4 started (bot: tutorial) 2014-08-28 01:32:14+0530 [scrapy] INFO: Optional features available: ssl, http11 2014-08-28 01:32:14+0530 [scrapy] INFO: Overridden settings: {'NEWSPIDER_MODULE' : 'tutorial.spiders', 'SPIDER_MODULES': ['tutorial.spiders'], 'BOT_NAME': 'tutor ial'} 2014-08-28 01:32:14+0530 [scrapy] INFO: Enabled extensions: LogStats, TelnetCons ole, CloseSpider, WebService, CoreStats, SpiderState Traceback (most recent call last): File "C:\Python27\lib\runpy.py", line 162, in _run_module_as_main "__main__", fname, loader, pkg_name) File "C:\Python27\lib\runpy.py", line 72, in _run_code exec code in run_globals File "c:\python27\Scripts\scrapy.exe\__main__.py", line 9, in <module> File "C:\Python27\lib\site-packages\scrapy\cmdline.py", line 143, in execute _run_print_help(parser, _run_command, cmd, args, opts) File "C:\Python27\lib\site-packages\scrapy\cmdline.py", line 89, in _run_print _help func(*a, **kw) File "C:\Python27\lib\site-packages\scrapy\cmdline.py", line 150, in _run_comm and cmd.run(args, opts) File "C:\Python27\lib\site-packages\scrapy\commands\crawl.py", line 60, in run self.crawler_process.start() File "C:\Python27\lib\site-packages\scrapy\crawler.py", line 92, in start if self.start_crawling(): File "C:\Python27\lib\site-packages\scrapy\crawler.py", line 124, in start_cra wling return self._start_crawler() is not None File "C:\Python27\lib\site-packages\scrapy\crawler.py", line 139, in _start_cr awler crawler.configure() File "C:\Python27\lib\site-packages\scrapy\crawler.py", line 47, in configure self.engine = ExecutionEngine(self, self._spider_closed) File "C:\Python27\lib\site-packages\scrapy\core\engine.py", line 64, in __init __ self.downloader = downloader_cls(crawler) File "C:\Python27\lib\site-packages\scrapy\core\downloader\__init__.py", line 73, in __init__ self.handlers = DownloadHandlers(crawler) File "C:\Python27\lib\site-packages\scrapy\core\downloader\handlers\__init__.p y", line 22, in __init__ cls = load_object(clspath) File "C:\Python27\lib\site-packages\scrapy\utils\misc.py", line 42, in load_ob ject raise ImportError("Error loading object '%s': %s" % (path, e)) ImportError: Error loading object 'scrapy.core.downloader.handlers.s3.S3Download Handler': No module named win32api C:\Python27\Scripts\tutorial> Thanks, Selvi Rathinam. -- You received this message because you are subscribed to the Google Groups "scrapy-users" group. To unsubscribe from this group and stop receiving emails from it, send an email to scrapy-users+unsubscr...@googlegroups.com. To post to this group, send email to scrapy-users@googlegroups.com. Visit this group at http://groups.google.com/group/scrapy-users. For more options, visit https://groups.google.com/d/optout.