DMARC for AI Agents — an open protocol for agent-to-agent trust verification in the agentic web.
Registry: api.airlock.ing — every verification routes through the central trust registry by default.
AI agents are rapidly gaining the ability to communicate with each other autonomously (via protocols like Google A2A and Anthropic MCP). There is no standard mechanism for verifying agent identity, authorization, or trustworthiness. The agent ecosystem is repeating the same mistake email made — building communication without authentication. Email took 20 years to bolt on SPF, DKIM, and DMARC after spam became an existential crisis. The Agentic Airlock builds the trust layer before the agent spam crisis hits.
A 5-phase cryptographic verification protocol with Ed25519 signing at every hop. Each agent interaction passes through:
Resolve → Handshake → Challenge → Verdict → Seal
95%+ of verifications complete in microseconds using pure cryptography. The semantic LLM challenge only fires for unknown agents — and only once per reputation tier.
┌─────────────────────────────────────────┐
│ Agentic Airlock │
│ │
Agent A ──────────► │ [Gateway] ──► EventBus │
(HandshakeRequest) │ │ │ │
│ │ ACK/NACK ▼ │
│ │ [Orchestrator] │
│ │ │ │
│ │ ┌─────┴──────┐ │
│ │ ▼ ▼ │
│ │ ReputationStore SemanticChallenge│
│ │ │ │ │
│ │ fast-path? ChallengeRequest │
│ │ │ → Agent A │
│ │ ▼ ▼ │
│ │ TrustVerdict (VERIFIED / │
│ │ REJECTED / DEFERRED) │
│ │ │ │
│ │ ▼ │
│ │ AirlockAttestation → Agent B │
└─────┴─────────────────────────────────── ┘
| # | Phase | What Happens |
|---|---|---|
| 1 | Resolve | Caller discovers the target agent's capabilities, DID, and endpoint status. The gateway looks up the agent registry and logs the event. |
| 2 | Handshake | Initiating agent presents a signed HandshakeRequest with its DID (did:key), intent, and a W3C Verifiable Credential. The gateway verifies the Ed25519 signature at transport time — invalid signatures are NACK'd instantly. |
| 3 | Challenge | If the agent's trust score is in the unknown zone (0.15–0.75), the orchestrator issues a ChallengeRequest — a semantic question about the agent's intended behaviour and capabilities. |
| 4 | Verdict | The orchestrator evaluates the challenge response (LLM-backed) and issues a signed TrustVerdict: VERIFIED, REJECTED, or DEFERRED. High-reputation agents skip phases 3 & 4 entirely (fast-path). |
| 5 | Seal | Both parties receive a signed SessionSeal containing the full verification trace, attestation, and updated trust score. The seal provides an auditable receipt for every interaction. |
pip install airlock-protocol
# Verify an agent in 7 lines
python -c "
from airlock import AirlockClient
client = AirlockClient() # defaults to api.airlock.ing
result = client.verify('did:key:z6MkhaXgBZDvotDkL5257faiztiGiC2QtKLGpbnnEGta2doK')
print(f'Verified: {result.verified}, Score: {result.trust_score}')
"# Verify an agent from the command line
airlock verify did:key:z6Mk...
# Start a local gateway for development
airlock serve
# Scaffold a new Airlock-protected project
airlock init# Clone and run locally
git clone https://github.com/airlock-protocol/airlock.git
cd airlock
pip install -e ".[dev]"
python demo/run_demo.py # 3-agent demo, no external services needed
python -m pytest tests/ -v # 313 testsfrom airlock import AirlockClient
# Default — routes through central Airlock registry (api.airlock.ing)
client = AirlockClient()
result = client.verify("did:key:z6Mk...")
if result.verified:
print(f"Trusted: {result.agent_name}, Score: {result.trust_score}")
# Self-hosted — point to your own gateway
client = AirlockClient(gateway_url="http://localhost:8000")
# Async support
result = await client.averify("did:key:z6Mk...")The npm workspace under sdks/typescript exposes the same REST operations via fetch (Node 18+). See sdks/typescript/README.md. Published PyPI name remains airlock-protocol (Python); the TS package is airlock-client on npm when released.
integrations/airlock-mcp is a stdio Model Context Protocol server that surfaces gateway tools (health, resolve, session, reputation, etc.) to MCP hosts. Build from repo root: npm install && npm run build:mcp.
When you publish: see RELEASING.md (PyPI OIDC, npm NPM_TOKEN, workflows).
- Docker Compose (gateway + Redis, persistent LanceDB volume): docs/deploy/docker.md
- Quick start: copy
.env.exampleto.env, setAIRLOCK_GATEWAY_SEED_HEX, thendocker compose up --build.
| Method | Endpoint | Description |
|---|---|---|
POST |
/resolve |
Look up an agent by DID and return its profile |
POST |
/handshake |
Submit a signed HandshakeRequest for verification |
POST |
/challenge-response |
Submit an agent's answer to a semantic challenge |
POST |
/register |
Register an AgentProfile (DID + capabilities + endpoint) |
POST |
/feedback |
Signed SignedFeedbackReport (Ed25519 + nonce); see SDKs |
POST |
/heartbeat |
Signed heartbeat (HeartbeatRequest with envelope + signature) |
GET |
/reputation/{did} |
Return the current trust score for an agent DID |
GET |
/session/{session_id} |
Poll session; use Authorization: Bearer with session_view_token from handshake ACK (or service token). Without auth in dev, trust_token is omitted. |
WS |
/ws/session/{session_id} |
Push session updates; same auth via Authorization or ?token= (session viewer JWT) |
GET |
/health |
Diagnostics (subsystems, queue depth, dead letters, uptime; HTTP 200 even if degraded) |
GET |
/live |
Process liveness (cheap; Docker HEALTHCHECK) |
GET |
/ready |
Readiness (HTTP 503 if deps not ready or shutting down) |
GET |
/metrics |
Prometheus text; requires AIRLOCK_SERVICE_TOKEN bearer when that env is set (always in AIRLOCK_ENV=production) |
POST |
/token/introspect |
Validate a trust JWT; requires gateway HS256 secret + service bearer when configured |
* |
/admin/* |
Optional ops API when AIRLOCK_ADMIN_TOKEN is set (Bearer) |
Public production: set AIRLOCK_ENV=production and the env vars documented in docs/deploy/docker.md (non-wildcard CORS, issuer allowlist, AIRLOCK_SERVICE_TOKEN, AIRLOCK_SESSION_VIEW_SECRET, etc.). LanceDB v1: use a single active writer or one replica with the LanceDB volume—see the deploy guide.
A2A routes under /a2a/* are documented in the gateway module; see airlock/gateway/a2a_routes.py.
New agents start at a neutral score of 0.50.
| Score Range | Routing Decision | Outcome |
|---|---|---|
≥ 0.75 |
Fast-path | VERIFIED immediately — no LLM challenge |
0.15 – 0.74 |
Semantic challenge | LLM evaluates the agent's intent |
≤ 0.15 |
Blacklist | REJECTED immediately |
| Verdict | Delta |
|---|---|
VERIFIED |
+0.05 / (1 + count × 0.1) (diminishing returns) |
REJECTED |
−0.15 (fixed penalty) |
DEFERRED |
−0.02 (small nudge — ambiguity is a signal) |
Scores decay toward neutral (0.50) over time using the standard radioactive decay formula:
decayed = 0.5 + (score − 0.5) × 2^(−elapsed_days / 30)
An agent that stops interacting gradually becomes "unknown" rather than "suspect" — matching real-world trust intuitions. The half-life is 30 days.
airlock-protocol/
├── airlock/
│ ├── config.py # Pydantic settings (env vars with AIRLOCK_ prefix)
│ ├── crypto/
│ │ ├── keys.py # Ed25519 KeyPair + did:key encoding/decoding
│ │ ├── signing.py # sign_model / verify_model + canonicalization
│ │ └── vc.py # W3C Verifiable Credential issue + validate
│ ├── engine/
│ │ ├── event_bus.py # Typed async EventBus (asyncio.Queue backed)
│ │ ├── orchestrator.py # LangGraph verification state machine (8 nodes)
│ │ └── state.py # SessionManager with TTL expiry
│ ├── gateway/
│ │ ├── app.py # FastAPI application factory + lifespan
│ │ ├── handlers.py # Request handlers (signature gate + event publish)
│ │ └── routes.py # FastAPI router + endpoint wiring
│ ├── reputation/
│ │ ├── scoring.py # Half-life decay + verdict delta computation
│ │ └── store.py # LanceDB-backed TrustScore persistence
│ ├── schemas/
│ │ ├── challenge.py # ChallengeRequest + ChallengeResponse
│ │ ├── envelope.py # MessageEnvelope, TransportAck, TransportNack
│ │ ├── events.py # VerificationEvent hierarchy (typed)
│ │ ├── handshake.py # HandshakeRequest + HandshakeResponse
│ │ ├── identity.py # AgentDID, AgentProfile, VerifiableCredential
│ │ ├── reputation.py # TrustScore schema
│ │ ├── session.py # VerificationSession + SessionSeal
│ │ └── verdict.py # TrustVerdict, AirlockAttestation, CheckResult
│ ├── sdk/
│ │ ├── client.py # AirlockClient (async httpx wrapper)
│ │ └── middleware.py # AirlockMiddleware (protect decorator)
│ └── semantic/
│ └── challenge.py # LLM-backed challenge generation + evaluation
├── integrations/
│ └── airlock-mcp/ # MCP stdio server (gateway tools)
├── sdks/
│ └── typescript/ # npm package `airlock-client` (HTTP + types)
├── examples/ # Agent scenarios + demos
└── tests/ # Pytest suite (gateway, engine, SDK, A2A, …)
| Principle | Implementation |
|---|---|
| PKI-first | All identities are did:key — DID documents derived from the Ed25519 public key, no registry required |
| Signed everything | Every message (HandshakeRequest, ChallengeRequest, ChallengeResponse, SessionSeal) carries an Ed25519 signature over its canonical JSON form |
| Challenge-response | Unknown agents face semantic questions that probe their stated capabilities — bad actors cannot fake plausible answers at scale |
| Event-driven | The gateway is a thin transport layer; all verification logic runs in an async EventBus + LangGraph state machine |
| Reputation with memory | Half-life decay means reputation is time-sensitive — a trusted agent that goes dark eventually becomes "unknown" again |
| Local-first | LanceDB is embedded (no server). The entire stack runs on a laptop: python demo/run_demo.py |
| A2A compatible | The HandshakeRequest schema is designed to wrap Google A2A message objects |
All settings can be configured via environment variables with the AIRLOCK_ prefix:
| Variable | Default | Description |
|---|---|---|
AIRLOCK_HOST |
0.0.0.0 |
Gateway bind address |
AIRLOCK_PORT |
8000 |
Gateway port |
AIRLOCK_SESSION_TTL |
180 |
Session expiry in seconds |
AIRLOCK_LANCEDB_PATH |
./data/reputation.lance |
Path to reputation database |
AIRLOCK_LITELLM_MODEL |
ollama/llama3 |
LLM model for semantic challenges |
AIRLOCK_LITELLM_API_BASE |
http://localhost:11434 |
LLM API endpoint |
Apache License 2.0. See LICENSE.
Shivdeep Singh (@shivdeep1) — airlock.ing