On-disk persistence for jobs and logs is generally considered a very good
thing.  For logs, it's the reason you want them.  I think it would be a
misdesign if invoking cancel erased all your logs, and your jobs (which is
there so you can restart a crawl after stopping it).

Why don't you just write a cronjob that cleans this up every night?

If you want it deleted immediately, just write a tiny API where a call to
your API generally wraps the scrapyd web interface?  In the case of cancel.
your script could delete the required fields after getting cancellation
confirmation from the scrapyd API.  A scrapyd web API middleware, if you
will.

On Wed, Oct 15, 2014 at 7:31 AM, Hugo Maugey <hugo.mau...@gmail.com> wrote:

> when I call :
>>
>
> curl http://localhost:6800/cancel.json -d project=backrub -d
> job=169a003c546f11e49251002590d225a2
>
> here is the answer :
>
> {"status": "ok", "prevstate": null}
>
> but the job (and its items and logs) are still there ...
>
> --
> You received this message because you are subscribed to the Google Groups
> "scrapy-users" group.
> To unsubscribe from this group and stop receiving emails from it, send an
> email to scrapy-users+unsubscr...@googlegroups.com.
> To post to this group, send email to scrapy-users@googlegroups.com.
> Visit this group at http://groups.google.com/group/scrapy-users.
> For more options, visit https://groups.google.com/d/optout.
>

-- 
You received this message because you are subscribed to the Google Groups 
"scrapy-users" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to scrapy-users+unsubscr...@googlegroups.com.
To post to this group, send email to scrapy-users@googlegroups.com.
Visit this group at http://groups.google.com/group/scrapy-users.
For more options, visit https://groups.google.com/d/optout.

Reply via email to