hi paul, pls ignore my question
On Thursday, 12 January 2017 16:07:43 UTC+5:30, vanam raghu wrote: > > Ok Paul i will do it , but i have one question when i try pip install > scrapy i think its installing 1.2.1 version i am not sure though why is it? > > Is pip not installing the latest version?? > > > > > On Thursday, 12 January 2017 15:53:11 UTC+5:30, Paul Tremberth wrote: >> >> Can you try updating scrapy to at least 1.2.2? >> >> On Thu, Jan 12, 2017 at 11:11 AM, vanam raghu <raghavendr...@gmail.com> >> wrote: >> >>> HI Paul >>> >>> Its >>> >>> Scrapy 1.2.1 >>> >>> >>> On Thursday, 12 January 2017 15:22:22 UTC+5:30, Paul Tremberth wrote: >>>> >>>> Hello Vanam, >>>> >>>> what version of Scrapy are you running? >>>> Scrapy 1.2.2 has a fix for a very similar error (if not the same): >>>> https://github.com/scrapy/scrapy/issues/2011 >>>> >>>> /Paul. >>>> >>>> On Thursday, January 12, 2017 at 10:43:14 AM UTC+1, vanam raghu wrote: >>>>> >>>>> While scheduling spider using scrapyd, i am seeing below error >>>>> >>>>> 2017-01-12 14:28:30 [scrapy] INFO: Spider opened >>>>> 2017-01-12 14:28:30 [manager] INFO: >>>>> -------------------------------------------------------------------------------- >>>>> 2017-01-12 14:28:30 [manager] INFO: Starting Frontier Manager... >>>>> 2017-01-12 14:28:30 [manager] INFO: Frontier Manager Started! >>>>> 2017-01-12 14:28:30 [manager] INFO: >>>>> -------------------------------------------------------------------------------- >>>>> 2017-01-12 14:28:30 >>>>> [frontera.contrib.scrapy.schedulers.FronteraScheduler] INFO: Starting >>>>> frontier >>>>> 2017-01-12 14:28:30 [scrapy] INFO: Closing spider (shutdown) >>>>> 2017-01-12 14:28:30 [twisted] CRITICAL: Unhandled error in Deferred: >>>>> 2017-01-12 14:28:30 [twisted] CRITICAL: >>>>> Traceback (most recent call last): >>>>> File >>>>> "/usr/local/lib/python2.7/dist-packages/twisted/internet/defer.py", line >>>>> 1258, in _inlineCallbacks >>>>> result = result.throwExceptionIntoGenerator(g) >>>>> File >>>>> "/usr/local/lib/python2.7/dist-packages/twisted/python/failure.py", line >>>>> 389, in throwExceptionIntoGenerator >>>>> return g.throw(self.type, self.value, self.tb) >>>>> File "/usr/local/lib/python2.7/dist-packages/scrapy/crawler.py", line >>>>> 87, in crawl >>>>> yield self.engine.close() >>>>> File "/usr/local/lib/python2.7/dist-packages/scrapy/core/engine.py", >>>>> line 100, in close >>>>> return self._close_all_spiders() >>>>> File "/usr/local/lib/python2.7/dist-packages/scrapy/core/engine.py", >>>>> line 340, in _close_all_spiders >>>>> dfds = [self.close_spider(s, reason='shutdown') for s in >>>>> self.open_spiders] >>>>> File "/usr/local/lib/python2.7/dist-packages/scrapy/core/engine.py", >>>>> line 298, in close_spider >>>>> dfd = slot.close() >>>>> File "/usr/local/lib/python2.7/dist-packages/scrapy/core/engine.py", >>>>> line 44, in close >>>>> self._maybe_fire_closing() >>>>> File "/usr/local/lib/python2.7/dist-packages/scrapy/core/engine.py", >>>>> line 51, in _maybe_fire_closing >>>>> self.heartbeat.stop() >>>>> File "/usr/local/lib/python2.7/dist-packages/twisted/internet/task.py", >>>>> line 202, in stop >>>>> assert self.running, ("Tried to stop a LoopingCall that was " >>>>> AssertionError: Tried to stop a LoopingCall that was not running. >>>>> >>>>> >>>>> >>>>> Can any body help me why i am getting this error?? >>>>> >>>>> -- >>> You received this message because you are subscribed to the Google >>> Groups "scrapy-users" group. >>> To unsubscribe from this group and stop receiving emails from it, send >>> an email to scrapy-users...@googlegroups.com. >>> To post to this group, send email to scrapy...@googlegroups.com. >>> Visit this group at https://groups.google.com/group/scrapy-users. >>> For more options, visit https://groups.google.com/d/optout. >>> >> >> -- You received this message because you are subscribed to the Google Groups "scrapy-users" group. To unsubscribe from this group and stop receiving emails from it, send an email to scrapy-users+unsubscr...@googlegroups.com. To post to this group, send email to scrapy-users@googlegroups.com. Visit this group at https://groups.google.com/group/scrapy-users. For more options, visit https://groups.google.com/d/optout.