missing caching for robots.txt ------------------------------ Key: DROIDS-105 URL: https://issues.apache.org/jira/browse/DROIDS-105 Project: Droids Issue Type: Improvement Components: core Reporter: Paul Rogalinski Attachments: CachingContentLoader.java
the current implementation of the HttpClient will not cache any requests to the robots.txt file. While using the CrawlingWorker this will result in 2 requests to the robots.txt (HEAD + GET) per crawled URL. So when crawling 3 URLs the target server would get 6 requests for the robots.txt. unfortunately the contentLoader is made final in HttpProtocol, so there is no possibility to replace it with a caching Protocol like that one you'll find in the attachment. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.