Hello. I'm writing a spider for a website that gathers some hyperlinks, then visits them and checks if something exists and returns the results into a text file. I have a for loop that yields requests, calling a parse2 function that checks the link and updates the text file.
evenselectorlist = response.css('table[id="result_table"] tr.even') for evenselector in evenselectorlist: relative = evenselector.css('a[title="Link"]::attr(href)').extract_first() yield scrapy.Request(response.urljoin(relative), callback=self.parse2,meta={'item':item},dont_filter=True) def parse2(self, response): #txt file stuff Is there a way to make the first parse function pause when the request is yielded? I would like to continue to do some stuff AFTER the new requests have ended. For example, I'd like to have a counter to see how many links have the information I want, which is available only after all the links have been visited. I hope you understand what I'm trying to say. Thank you! -- You received this message because you are subscribed to the Google Groups "scrapy-users" group. To unsubscribe from this group and stop receiving emails from it, send an email to scrapy-users+unsubscr...@googlegroups.com. To post to this group, send email to scrapy-users@googlegroups.com. Visit this group at https://groups.google.com/group/scrapy-users. For more options, visit https://groups.google.com/d/optout.