Skip to content

Add robots.txt checking option #1

@ArkNill

Description

@ArkNill

Feature Request

Currently markgrab does not check robots.txt before fetching. Add an optional respect_robots=True parameter that:

  1. Fetches and parses robots.txt for the target domain
  2. Checks if the URL path is allowed for the configured user agent
  3. Raises RobotsDisallowed or silently skips if disallowed

This should be opt-in (default False) to maintain backward compatibility.

Motivation

Legal compliance for production deployments. Currently documented in Disclaimer but not enforced.

Metadata

Metadata

Assignees

No one assigned

    Labels

    enhancementNew feature or request

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions