Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Respect robots.txt when crawling when set as True #42

Closed
indrajithi opened this issue Jun 18, 2024 · 1 comment · Fixed by #45
Closed

Respect robots.txt when crawling when set as True #42

indrajithi opened this issue Jun 18, 2024 · 1 comment · Fixed by #45
Assignees

Comments

@indrajithi
Copy link
Collaborator

indrajithi commented Jun 18, 2024

  • Option to set respect_robots_txt (Default should be True because of legal obligation in some jurisdictions)
  • Fetch and Parse robots.txt (urllib.robotparser will helps in parsing robots.txt)
  • Create crawl rule per domain
  • Check URL permissions before crawling a URL
  • Make sure it works when concurrent workers are fetching different domains
  • Use the rules provided in robots.txt to fetch. (eg: use robots.txt crawl-delay if present. Check the rule before crawling a path)
@indrajithi indrajithi changed the title Respect robots.txt when crawling Respect robots.txt when crawling when set as True Jun 18, 2024
@Mews
Copy link
Collaborator

Mews commented Jun 18, 2024

@indrajithi I started working on this already, assign me when you can :)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging a pull request may close this issue.

2 participants