(feat) Add support for robots.txt by default #668
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
Related Issues
N/A
Proposed Changes:
This PR adds support for robots.txt by default. This PR uses the Protego lbrary to parse robots.txt for source files in the source.download method. All subsequent downloads have to pass robots.txt to go through. If robots.txt is not found, all requests should pass.
Whenever your bot is disallowed for the source url, we throw a RobotsException. If your bot is disallowed for some article url, we just print a warning and keep going, no exception.
This should help people avoid being banned by news sites for scraping.
If you still want to ignore robots.txt there is a new source.config option called dont_obey_robotstxt that you can set to true to ignore robots.txt. This option is False by default.
How did you test it?
I added a unit test for robots.txt with a user agent that people like to disallow. I regression tested using the existing tests.
Notes for the reviewer
Checklist
fix:
,feat:
,build:
,chore:
,ci:
,docs:
,style:
,refactor:
,perf:
,test:
.