Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Speculation on how sub-googlebots work, based on public documentation #2

wants to merge 7 commits into
base: master
Choose a base branch


Copy link

@willcritchlow willcritchlow commented Nov 12, 2019

If we pass in a comma-separated list of two user agents (more specific first, e.g. googlebot-image,googlebot) then run newly-exposed method that returns a response based on whether the first user agent is allowed / disallowed by a specific rule, and if not, falls back to a regular check on the second user agent.

Will Critchlow and others added 7 commits Nov 12, 2019
First should be the specific user agent, and second should be the general user agent whose ruleset the crawler should obey iif there are no specific rules targeting the first user agent
Expose tuple functionality in the interface to robots and call from the wrapper if two user agents passed in the string comma-separated.

Add test cases.
and documentation
One of the ways listed in the open source project for compiling and testing the project is using bazel - which creates directories that should be ignored by git as we never want to check them in
Ignore bazel build directories
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
None yet
None yet
1 participant