Hi,
I'm trying to use scrapy to do a broad crawl. What I'm doing is to follow 
every link I find on every page, if the domain matches a rule. I feed the 
Spider with a few public directories of websites, and I use very high 
concurrency. At the beginning it's fast (first minute: 1381 pages/min), but 
then the speed decreases every minute down to 50 pages/min after 6 minutes. 
Then it's slow and stable ;)

CPU, memory, network and disk usages are very low, after the initial peak.

I noticed more than 100k requests in the queue after a few minutes, and 
just a few thoundands crawled pages.

I tried different settings for CONCURRENT_REQUESTS, 
CONCURRENT_REQUESTS_PER_IP and added more start_urls, but it only increased 
the initial peak.

Is there something wrong with my settings? 
https://gist.github.com/vad/3c3859ee17c07bcb3636

In the gist I also put the stack trace i see in every thread (i used the 
debugging middleware).

Regards

-- 
You received this message because you are subscribed to the Google Groups 
"scrapy-users" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to scrapy-users+unsubscr...@googlegroups.com.
To post to this group, send email to scrapy-users@googlegroups.com.
Visit this group at http://groups.google.com/group/scrapy-users.
For more options, visit https://groups.google.com/d/optout.

Reply via email to