Skip to content

WhiteFinSec/warden

Use this GitHub action with your project
Add this Action to an existing workflow or create a new one
View on Marketplace

Repository files navigation

Warden — AI Agent Governance Scanner

PyPI version License: MIT Python 3.10+

Open-source, local-only CLI scanner that evaluates AI agent governance posture across 12 scan layers and 17 dimensions. Scans code patterns, MCP configs, infrastructure, secrets, agent architecture, dependencies, audit compliance, CI/CD pipelines, IaC security, framework-specific governance, multi-language code, and cloud AI services. No data leaves the machine.

Website: whitefin.ai · PyPI: warden-ai · GitHub Action: Marketplace listing

For security teams / CISOs: zero local install required. Scan any GitHub repo by adding the Warden GitHub Action to its workflows — runs on GitHub's infrastructure, posts findings to your Code Scanning tab.

Install

Pick the tier that matches your environment. Zero Python required for Tier 1 — scanning runs on GitHub's runners.

Tier 1 — Zero install (GitHub Action)

For security teams, CISOs, and anyone who wants governance scoring in CI without touching local Python. Drop a workflow file into any repo:

# .github/workflows/warden.yml
name: Warden governance scan
on: [push, pull_request]
permissions:
  contents: read
  security-events: write   # for SARIF upload to Code Scanning
jobs:
  warden:
    runs-on: ubuntu-latest
    steps:
      - uses: actions/checkout@v4
      - uses: whitefinsec/warden@v1
        with:
          path: .
          # Optional: fail the build on poor posture
          # min-score: 60
          # fail-on-level: at_risk

Findings show up in your repo's Security → Code scanning tab, scoped per file and line. Full input/output reference in the GitHub Action — Reference section below.

Tier 2 — One-shot local scan

For a quick local audit. No persistent install — uvx runs Warden in a throwaway environment.

# Install uv first if you don't have it (one-time, ~10 MB):
#   curl -LsSf https://astral.sh/uv/install.sh | sh        # macOS / Linux
#   powershell -c "irm https://astral.sh/uv/install.ps1 | iex"   # Windows

uvx --from warden-ai warden scan /path/to/your-agent-project

Note: the package is warden-ai but the CLI is warden — that's why the --from flag is needed for uvx.

Tier 3 — Persistent CLI install

For repeated use or scripting. Recommended for developers.

# pipx (recommended — isolated env, `warden` on PATH everywhere)
pipx install warden-ai

# Or pip into your user site
pip install --user warden-ai

# Then:
warden scan /path/to/your-agent-project
warden --version

Optional extras:

pipx install 'warden-ai[pdf]'   # adds `--format pdf` (weasyprint)

From zero to governance score in under 60 seconds.

Windows notes

  • If pip install or uv tool install hangs silently for minutes, add Defender exclusions for your Python and uv directories (admin PowerShell):
    Add-MpPreference -ExclusionPath "$env:APPDATA\Python"
    Add-MpPreference -ExclusionPath "$env:LOCALAPPDATA\uv"
    Add-MpPreference -ExclusionPath "$env:APPDATA\uv"
    Add-MpPreference -ExclusionProcess "python.exe"
    Add-MpPreference -ExclusionProcess "py.exe"
    Add-MpPreference -ExclusionProcess "uv.exe"
  • If uvx hangs after "Acquired shared lock", a previous run left a stale lock. Clear it:
    Get-Process uv,uvx,uvw,python -ErrorAction SilentlyContinue | Stop-Process -Force
    Remove-Item "$env:LOCALAPPDATA\uv\cache\.lock" -Force -ErrorAction SilentlyContinue

HTML Report

Warden generates a self-contained HTML report with interactive score breakdown, actionable recommendations, and a comparison card — works offline and in air-gapped environments.

Warden HTML Report

What It Does

Warden scores your AI agent project across 17 governance dimensions (out of 235 raw points, normalized to /100):

Group Dimensions
Core Governance (100 pts) Tool Inventory, Risk Detection, Policy Coverage, Credential Management, Log Hygiene, Framework Coverage
Advanced Controls (50 pts) Human-in-the-Loop, Agent Identity, Threat Detection
Ecosystem (55 pts) Prompt Security, Cloud/Platform, LLM Observability, Data Recovery, Compliance Maturity
Unique Capabilities (30 pts) Post-Exec Verification, Data Flow Governance, Adversarial Resilience

Score Levels

Score Level Meaning
>= 80 GOVERNED Comprehensive agent governance in place
>= 60 PARTIAL Significant coverage with material gaps
>= 33 AT_RISK Some controls exist but major blind spots
< 33 UNGOVERNED Minimal or no agent governance

CLI Commands

# Scan a project (generates HTML + JSON + SARIF reports)
warden scan .
warden scan /path/to/project --format json
warden scan /path/to/project --format sarif
warden scan /path/to/project --format pdf       # requires pip install 'warden-ai[pdf]'
warden scan /path/to/project --output-dir /path/to/reports

# Skip specific layers
warden scan . --skip secrets,deps

# Run only specific layers
warden scan . --only code,mcp,cloud

# CI mode: exit code reflects governance level
warden scan . --ci                    # 0=governed, 1=partial, 2=at_risk, 3=ungoverned
warden scan . --min-score 60          # exit 1 if score < 60

# Baseline: track only new findings (brownfield adoption)
warden baseline .                     # saves .warden-baseline.json
warden scan . --baseline .warden-baseline.json  # shows only NEW findings

# Compare two reports
warden diff before.json after.json    # score delta, new/resolved findings

# Auto-fix common findings
warden fix . --dry-run                # preview fixes
warden fix .                          # apply fixes

# View the scoring methodology
warden methodology

# See the market leaderboard (20 vendors x 17 dimensions)
warden leaderboard

Config File (.warden.toml)

Warden reads project-level defaults from .warden.toml or a [tool.warden] table in pyproject.toml. Values apply when the matching CLI flag is left at its default — explicit flags always win.

# .warden.toml — checked into your repo
format = "all"
output_dir = "reports"
skip = ["secrets", "deps"]
only = []
min_score = 60
baseline = ".warden-baseline.json"
ci = true

Or alongside other tooling in pyproject.toml:

[tool.warden]
format = "sarif"
min_score = 70
skip = ["multilang"]

Warden searches upward from the scan path until it finds a config file or hits a VCS root (.git, .hg, .svn). Paths like output_dir and baseline are resolved relative to the config file. Pass --no-config to ignore any discovered config.

GitHub Action — Reference

Full input/output interface for the Tier 1 GitHub Action above. Every push and PR scores governance posture and publishes findings to Code Scanning.

# .github/workflows/warden.yml
name: Warden governance scan

on:
  push:
    branches: [main]
  pull_request:
    branches: [main]

permissions:
  contents: read
  security-events: write   # required for SARIF upload

jobs:
  warden:
    runs-on: ubuntu-latest
    steps:
      - uses: actions/checkout@v4
      - uses: whitefinsec/warden@v1
        with:
          path: .
          # Optional gates:
          # min-score: 60          # fail the build if score < 60
          # fail-on-level: at_risk # fail if posture is AT_RISK or worse

Key action inputs: path, format (json/html/sarif/all), min-score, fail-on-level, skip, only, baseline, upload-sarif, warden-version, python-version. Outputs: score, raw-score, level, findings-count, critical-count, report-json, report-html, report-sarif. See action.yml for the full interface and .github/workflows/warden-example.yml.sample for a full example workflow.

Layer Keys for --skip / --only

Key Layer
code Code Patterns (Python AST + JS/TS regex)
mcp MCP Server Configs
infra Infrastructure (Docker, K8s)
secrets Secrets & Credentials
agent Agent Architecture
deps Supply Chain / Dependencies
audit Audit & Compliance
cicd CI/CD Governance
iac IaC Security (Terraform, Pulumi, CloudFormation)
frameworks Framework-Specific Governance
multilang Multi-Language Governance (Go, Rust, Java)
cloud Cloud AI Governance (AWS, Azure, GCP)

12 Scan Layers

  1. Code Patterns — AST-based Python + regex JS/TS analysis (unprotected LLM calls, agent loops, unrestricted tool access)
  2. MCP Servers — Config file analysis (write tools without auth, missing schemas, non-TLS transport)
  3. Infrastructure — Dockerfile, docker-compose, K8s manifests (root containers, exposed secrets, missing healthchecks)
  4. Secrets — 15+ credential patterns with value masking (OpenAI, Anthropic, AWS, GitHub, Stripe, etc.)
  5. Agent Architecture — Agent class analysis (no permissions, no cost tracking, unlimited sub-agent spawning)
  6. Supply Chain — Dependency analysis (unpinned AI packages, typosquat detection via Levenshtein distance)
  7. Audit & Compliance — Audit logging, structured logging, retention policies, compliance framework mapping
  8. CI/CD Governance — GitHub Actions analysis (missing approvals, exposed secrets, no branch protection, CODEOWNERS)
  9. IaC Security — Terraform, Pulumi, and CloudFormation analysis (unencrypted storage, open security groups, IAM wildcards, missing remote backend)
  10. Framework Governance — LangChain callbacks, CrewAI guardrails, AutoGen sandboxing, LlamaIndex limits
  11. Multi-Language Governance — Go (context timeouts, unsafe exec), Rust (unsafe blocks, .unwrap() on API calls), Java (Spring AI @Tool auth, audit logging)
  12. Cloud AI Governance — AWS Bedrock guardrails, Azure AI Content Safety, GCP Vertex AI safety settings, managed identity vs hardcoded keys

Plus D17: Adversarial Resilience — 8 sub-checks based on Google DeepMind's "AI Agent Traps" paper (Franklin et al., March 2026).

Scoring Integrity

Warden v1.5+ includes 6 anti-inflation mechanisms to prevent score gaming:

  • Strong/weak pattern tiers — generic matches (e.g., import logging) score 1 point; governance-specific patterns (e.g., audit_log_tamper_proof) score 3
  • Co-occurrence requirements — dimensions like D3 (Policy) and D11 (Cloud/Platform) require 3+ distinct patterns to score, preventing single-keyword inflation
  • Boolean dimension scoring — each dimension scores from code patterns OR absence, never both
  • CRITICAL finding deductions — each CRITICAL finding deducts 2 points (capped at 60% of earned score)
  • MCP absence-vs-compliance fix — "no tools found = no violations" no longer counts as compliant; only inline tool definitions earn credit
  • Positive-signal scoring — clean dependencies and zero secrets earn modest credit (1-3 pts), not full dimension scores; real points require active governance patterns (secrets managers, compliance frameworks, lockfiles)

HTML Report Features

The HTML report is fully self-contained (no CDN, no external fonts, no network requests):

  • Score gauge with per-dimension breakdown bars
  • Summary grid — MCP-focused when MCP tools detected, findings-focused otherwise
  • Discovered tools — MCP tool inventory with risk classification (destructive, financial, exfiltration, write-access, read-only)
  • Governance detection — which governance layers were found in your codebase
  • Recommendations — prioritized remediation steps mapped to compliance frameworks
  • Comparison card — side-by-side score projection with biggest gap dimensions
  • Competitor detection — identifies 20 governance/security tools in your codebase (shown only when detected, requires 2+ signals)
  • Email form — optional report delivery (score metadata only, never source code or secrets)

Output Formats

Format File Description
HTML warden_report.html Self-contained dark-theme report with SVG gauge, expandable findings, benchmark bars
JSON warden_report.json Machine-readable with scoring_version field for CI/CD integration
SARIF warden_report.sarif GitHub Code Scanning compatible — native PR annotations
CLI stdout Colorized terminal output with per-layer timing and progress bars

Language Support

Language Code Patterns Secrets Dependencies Framework-Specific Cloud AI
Python AST Yes pip/poetry/uv LangChain, CrewAI, AutoGen, LlamaIndex Bedrock, Azure AI, Vertex AI
JavaScript/TypeScript Regex Yes npm/yarn/pnpm
Go Regex Yes go.mod context, exec, rate limiting
Rust Regex Yes Cargo.toml tracing, tokio, unsafe blocks
Java Regex Yes Maven/Gradle Spring AI, Spring Security
Terraform HCL regex Provider versions
Pulumi Via TS/PY
CloudFormation YAML/JSON regex

Architecture Constraints

  1. Zero network access — Scanners never import httpx/requests/urllib. CI-enforced.
  2. Zero WhiteFin imports — Standalone package with no internal dependencies. CI-enforced.
  3. Secrets never stored — Only file, line, pattern name, and masked preview (first 3 + last 4 chars).
  4. HTML report self-contained — No CDN, no Google Fonts. Works in air-gapped environments.
  5. 2 runtime dependencies — click + rich. Nothing else.

Development

# With uv (recommended)
uv sync --extra dev
uv run pytest tests/ -v

# With pip
python -m venv .venv
source .venv/bin/activate  # or .venv\Scripts\activate on Windows
pip install -e ".[dev]"
pytest tests/ -v

Known Limitations

  • Static analysis: Warden detects governance patterns, not enforcement. High score = controls present, not proven correct.
  • Framework vocabulary: Scoring is optimized for recognized AI frameworks. Custom frameworks may score lower despite equivalent governance.
  • IaC depth: Terraform has the deepest analysis. Pulumi and CloudFormation checks are regex-based heuristics.
  • Multi-language AST: Go/Rust/Java analysis uses regex, not AST parsing. Fewer patterns detected than Python.
  • Local filesystem scope: Warden scans files on disk, including gitignored files. Secrets in .env files are flagged even if not committed.

See SCORING.md for full methodology details.

License

MIT

Research Citation

Adversarial resilience dimension (D17) cites:

Franklin, Tomasev, Jacobs, Leibo, Osindero. "AI Agent Traps." Google DeepMind, March 2026.

Every D17 finding maps to EU AI Act articles, OWASP LLM Top 10, and MITRE ATLAS techniques.

About

AI Agent Governance Scanner — 17-dimension scoring across 7 scan layers. Local-only, privacy-first.

Resources

Contributing

Stars

Watchers

Forks

Packages

 
 
 

Contributors