Hello, Storage is handled by scrapy itself. You should include a pipeline for mongodb in your project.
You can schedule your spiders with curl and cron since you need a simple weekly schedule. scrapyd looks in several places for the configuration file. If you use the latest version, ~/.scrapyd.conf is one of them. scrapyd is no longer packaged for debian/ubuntu. I was thinking of including manual install instructions in the docs. Sharing information on your experience can help improve scrapyd so if you encounter other problems, don't hesitate to reply to this email. On Wednesday, 15 March 2017 13:51:32 UTC+2, odin wrote: > > First time I installed *scrapyd *in Ubuntu 14.04**, I didn't use the > generic way. > Using apt-get, my *scrapyd *was considered a service that can be started > and have (log/config/dbs...) dependencies but the *scrapy *version was > very outdated. > > So I installed *scrapyd *with pip in *virtualenv*. > Although it is up to date I can't start *scrapyd *as a service and I > can't find any dependencies. > Where do I create the *Configuration *file to link (eggs/dbs/items/log) > dependencies ? > > I have more than 10 spiders. Using a *remote Ubuntu server*, I want each > spider to scrap *periodically *(once a weak for example) and send the > data in a *mangodb *database. Most of the spiders don't have to scrap > simultaneously. > > What is the *best approach* to run *scrapyd *as a service and run spiders > periodically in my Ubuntu server? > -- You received this message because you are subscribed to the Google Groups "scrapy-users" group. To unsubscribe from this group and stop receiving emails from it, send an email to scrapy-users+unsubscr...@googlegroups.com. To post to this group, send email to scrapy-users@googlegroups.com. Visit this group at https://groups.google.com/group/scrapy-users. For more options, visit https://groups.google.com/d/optout.