Hello Vanam, what version of Scrapy are you running? Scrapy 1.2.2 has a fix for a very similar error (if not the same): https://github.com/scrapy/scrapy/issues/2011
/Paul. On Thursday, January 12, 2017 at 10:43:14 AM UTC+1, vanam raghu wrote: > > While scheduling spider using scrapyd, i am seeing below error > > 2017-01-12 14:28:30 [scrapy] INFO: Spider opened > 2017-01-12 14:28:30 [manager] INFO: > -------------------------------------------------------------------------------- > 2017-01-12 14:28:30 [manager] INFO: Starting Frontier Manager... > 2017-01-12 14:28:30 [manager] INFO: Frontier Manager Started! > 2017-01-12 14:28:30 [manager] INFO: > -------------------------------------------------------------------------------- > 2017-01-12 14:28:30 [frontera.contrib.scrapy.schedulers.FronteraScheduler] > INFO: Starting frontier > 2017-01-12 14:28:30 [scrapy] INFO: Closing spider (shutdown) > 2017-01-12 14:28:30 [twisted] CRITICAL: Unhandled error in Deferred: > 2017-01-12 14:28:30 [twisted] CRITICAL: > Traceback (most recent call last): > File "/usr/local/lib/python2.7/dist-packages/twisted/internet/defer.py", > line 1258, in _inlineCallbacks > result = result.throwExceptionIntoGenerator(g) > File "/usr/local/lib/python2.7/dist-packages/twisted/python/failure.py", > line 389, in throwExceptionIntoGenerator > return g.throw(self.type, self.value, self.tb) > File "/usr/local/lib/python2.7/dist-packages/scrapy/crawler.py", line 87, > in crawl > yield self.engine.close() > File "/usr/local/lib/python2.7/dist-packages/scrapy/core/engine.py", line > 100, in close > return self._close_all_spiders() > File "/usr/local/lib/python2.7/dist-packages/scrapy/core/engine.py", line > 340, in _close_all_spiders > dfds = [self.close_spider(s, reason='shutdown') for s in > self.open_spiders] > File "/usr/local/lib/python2.7/dist-packages/scrapy/core/engine.py", line > 298, in close_spider > dfd = slot.close() > File "/usr/local/lib/python2.7/dist-packages/scrapy/core/engine.py", line > 44, in close > self._maybe_fire_closing() > File "/usr/local/lib/python2.7/dist-packages/scrapy/core/engine.py", line > 51, in _maybe_fire_closing > self.heartbeat.stop() > File "/usr/local/lib/python2.7/dist-packages/twisted/internet/task.py", > line 202, in stop > assert self.running, ("Tried to stop a LoopingCall that was " > AssertionError: Tried to stop a LoopingCall that was not running. > > > > Can any body help me why i am getting this error?? > > -- You received this message because you are subscribed to the Google Groups "scrapy-users" group. To unsubscribe from this group and stop receiving emails from it, send an email to scrapy-users+unsubscr...@googlegroups.com. To post to this group, send email to scrapy-users@googlegroups.com. Visit this group at https://groups.google.com/group/scrapy-users. For more options, visit https://groups.google.com/d/optout.