Just answered my own questions:

(1) Do NOT call get_project_settings() in spider's __init__ method because: 
this will retrieve only the settings in settings.py, not any custom spider 
or command line settings.
(2) In order to pass arguments into the spider's __init__ method from the 
from_crawler method, make sure both methods accept *args and **kwargs as 
input. Then you can do something like this in from_crawler:

my_setting = crawler.settings.get('MY_SETTING')

kwargs['foo'] = my_setting

obj = super(MySpider, self).from_crawler(crawler, *args, **kwargs)

return obj

On Wednesday, April 5, 2017 at 3:33:55 PM UTC-7, Andre King wrote:
>
> So far, I've read that settings should be gotten in the from_crawler 
> method and passed in as args or kwargs to the spider __init__ method.
>
> I wasn't able to figure out how to do this, so my solution has been to 
> simply call get_project_settings() in the spider's __init__ method instead 
> and go from there.
>
> Is there anything wrong with this approach? If so, can someone provide an 
> example as to how to pass settings to __init__ from from_crawler?
>
> Thanks,
> Andre
>

-- 
You received this message because you are subscribed to the Google Groups 
"scrapy-users" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to scrapy-users+unsubscr...@googlegroups.com.
To post to this group, send email to scrapy-users@googlegroups.com.
Visit this group at https://groups.google.com/group/scrapy-users.
For more options, visit https://groups.google.com/d/optout.

Reply via email to