If I have a *really *slow pipeline, like let's say I'm writing out items to a database on a remote server that is really slow, what would happen? Would the items just stack up in memory until finally they are processed (meaning my only problem might be memory) or would Scrapy crawling of pages halt because of this too?
I'm thinking that when an item is passed to `process_item` method of the pipeline, Scrapy just carries on to the next request regardless of what happens in the pipeline? I'm using a MS-SQL writer pipeline based on dirbot-mysql <https://github.com/darkrho/dirbot-mysql>, but adapted to MS-SQL. I'm trying to understand the real advantages of using twisted adbapi though. I understand it will speed up the writing of items to the db: since asynchronously it will switch between connections in the pool; if a connection starts blocking, jump to another that isn't blocking. But this is just the writing of items phase right? If the pipeline isn't blocking the crawl then so what? -- You received this message because you are subscribed to the Google Groups "scrapy-users" group. To unsubscribe from this group and stop receiving emails from it, send an email to scrapy-users+unsubscr...@googlegroups.com. To post to this group, send email to scrapy-users@googlegroups.com. Visit this group at http://groups.google.com/group/scrapy-users. For more options, visit https://groups.google.com/d/optout.