[MRG+1] Enable robots.txt handling by default for new projects. #1724
Conversation
For backwards compatibility reasons the default value is not changed.
Current coverage is
|
Neat idea. |
@eliasdorneles do you mean a deprecation warning with ROBOTSTXT_OBEY is not set? It could be inconvenient for CrawlerProcess users because Changing What about changing default values in future Scrapy 2.0? |
Hmm, right. |
|
[MRG+1] Enable robots.txt handling by default for new projects.
I also agree that we can make it default setting in v2.0 |
I have the impression that I referenced this already but can't find it. |
@Digenis do you suggest to enable or disable robots.txt handling when user agent is overridden? We encourage users to override user agent in generated setting.py (near the top), and provide an example: |
Not permanently tied to robots. Just for an extra release in deprecation warnings. |
A proposed fix for #1668.
For backwards compatibility reasons the default value is not changed: settings for existing projects or settings used for CrawlerProcess won't change.