-
Notifications
You must be signed in to change notification settings - Fork 6.5k
[FEATURE] EU AI Act compliance: audit logging & human oversight for autonomous agent crews #4554
Description
Feature Area
Core functionality / Agent capabilities
Is your feature request related to an existing bug?
NA — This is a proactive compliance feature request.
Describe the solution you'd like
The EU AI Act (Regulation 2024/1689) enters enforcement in August 2026 and places specific requirements on autonomous AI agent systems — exactly the kind CrewAI enables. Key articles relevant to multi-agent orchestration:
- Article 9 (Risk Management): Agent crews performing high-risk tasks (healthcare, finance, legal) need documented risk assessment
- Article 13 (Transparency): Users must understand which agent made which decision, with what tools, and why
- Article 14 (Human Oversight): Autonomous crews need a mechanism for human intervention/override at critical decision points
- Article 17 (Quality Management): Agent outputs should be auditable with quality metrics
Proposed additions:
- Audit trail per crew execution: Structured log of each agent's reasoning, tool calls, delegations, and outputs — enabling post-hoc compliance review
- Risk classification helper: Utility to classify a crew's use case against EU AI Act risk tiers (minimal / limited / high / unacceptable)
- Human-in-the-loop hooks: Built-in support for mandatory human approval before certain agent actions (already partially exists via
human_input=True, but could be formalized for compliance) - Compliance metadata in crew output: Include provenance data (which model, which tools, which data sources) in the crew's final result
Why this matters now
- The EU AI Act applies to any AI system deployed in or affecting EU users, regardless of where the developer is based
- Multi-agent systems are likely to be classified as high-risk under Annex III when used in critical domains
- Adding compliance features now gives CrewAI a competitive advantage as the first agent framework with built-in EU AI Act support
Describe alternatives you've considered
- External compliance wrapper: Users build their own audit logging around CrewAI — but this is fragmented and error-prone
- Third-party compliance tools: Tools like arkforge.fr/mcp-eu-ai-act can scan codebases for AI framework usage and flag compliance gaps, but framework-level support is more robust
- Documentation only: A compliance guide without code changes — helpful but insufficient for automated auditing
Additional context
I work on EU AI Act compliance tooling and have been tracking how major AI frameworks are preparing. CrewAI's agent orchestration model is powerful, but the lack of built-in audit trails makes it harder for enterprise users to adopt in regulated industries.
Happy to contribute a PR for the audit logging component if there's interest. The transparency requirements (Article 13) map naturally to CrewAI's existing callback system.
Related regulation: EU AI Act full text