Re: [dspace-tech] High traffic / DDoS / fail2ban

2023-01-19 Thread Edmund Balnaves
A robots.txt file can help with many spiders, along with a link to the dspace sitemap. Sitemap: /jspui/sitemap The robots.txt file can include Crawl-delay: 10 and it is useful to disallow the search and browse links - eg Disallow: /jspui/simple-search May robots get lost in circling

Re: [dspace-tech] High traffic / DDoS / fail2ban

2023-01-19 Thread Mark H. Wood
On Thu, Jan 19, 2023 at 11:50:03AM +0100, Florian Wille wrote: > my DSpace (6.3) Site usually gets around 10k/h requests. This is handled > quite well. But sometimes there are multiple > bots/crawlers/spiders/indexers/harvester/whatevers throwing each up to > 15k/h request at me at the same

[dspace-tech] High traffic / DDoS / fail2ban

2023-01-19 Thread Florian Wille
Hey There, my DSpace (6.3) Site usually gets around 10k/h requests. This is handled quite well. But sometimes there are multiple bots/crawlers/spiders/indexers/harvester/whatevers throwing each up to 15k/h request at me at the same time and that on top of my 10k/h standart traffic. This my