The AI-first multi-model database.
Tables. Documents. Graphs. Vectors. KV. One engine. Ask it anything.
Important
In RedDB, a collection is the named logical container for data. Tables, documents, key-value, graphs,
vectors, time-series, and queues are the user-facing models or semantics you use on top of collections.
A collection is not a separate hierarchy layer that contains multiple tables and documents beneath it.
Instead, users can be a collection used as a table, events can be a collection used for documents,
and config can be a collection used as KV. Some models can also coexist in the same collection.
ASK 'who owns passport AB1234567 and what services do they use?'One command. RedDB searches across tables, graphs, vectors, documents, and key-value stores -- builds context -- calls an LLM -- returns a natural-language answer. No pipelines. No glue code. No other database does this.
Stop running Postgres + Neo4j + Pinecone + Redis + Mongo + InfluxDB + RabbitMQ. RedDB unifies them.
The key mental model is simple: the model is how you work with the data, and the collection is where that data lives.
-- Relational rows
INSERT INTO users (name, email) VALUES ('Alice', 'alice@co.com')
-- JSON documents
INSERT INTO logs DOCUMENT (body) VALUES ('{"level":"info","msg":"login"}')
-- Graph edges
INSERT INTO network EDGE (label, from_rid, to_rid) VALUES ('CONNECTS', 102, 103)
-- Vector similarity search
SEARCH SIMILAR TEXT 'anomaly detected' COLLECTION events
-- Key-value
KV PUT config.theme = 'dark'
-- Time-series metrics (with retention & downsampling)
CREATE TIMESERIES cpu_metrics RETENTION 90 d
INSERT INTO cpu_metrics (metric, value, tags) VALUES ('cpu.idle', 95.2, {"host":"srv1"})
-- Hypertables + partition TTL + continuous aggregates (logs / events / telemetry)
CREATE HYPERTABLE access_log TIME_COLUMN ts CHUNK_INTERVAL '1d' TTL '90d'
-- Append-only tables (audit, ledger, immutable events)
CREATE TABLE audit_log (id BIGINT, action TEXT) APPEND ONLY
-- Message queues (FIFO, priority, consumer groups)
CREATE QUEUE tasks MAX_SIZE 10000
QUEUE PUSH tasks {"job":"process","id":123}
QUEUE POP tasksSame file. Same engine. Same query language.
Want to use RedDB as your log store? Start with the Logs Quickstart or the full Using RedDB for Logs guide.
-- Semantic search without managing vectors yourself
SEARCH SIMILAR TEXT 'suspicious login' COLLECTION logs USING groq
-- Auto-embed on insert -- vectors are created for you
INSERT INTO articles (title, body) VALUES ('AI Safety', 'Alignment research...')
WITH AUTO EMBED (body) USING openai
-- Context search: find everything related to an entity across all models
SEARCH CONTEXT '192.168.1.1' FIELD ip DEPTH 2
-- Ask questions in plain English
ASK 'what vulnerabilities affect host 10.0.0.1?' USING anthropicRedDB retrieves context from every data model, feeds it to the LLM, and gives you a grounded answer. RAG built into the database layer.
Swap providers with a keyword. No code changes.
| Provider | Keyword | API Key | ASK / Prompt | Embeddings |
|---|---|---|---|---|
| OpenAI | openai |
Yes | ✅ | ✅ |
| Anthropic | anthropic |
Yes | ✅ | rejected |
| Groq | groq |
Yes | ✅ | ✅ |
| OpenRouter | openrouter |
Yes | ✅ | ✅ |
| Together | together |
Yes | ✅ | ✅ |
| Venice | venice |
Yes | ✅ | ✅ |
| DeepSeek | deepseek |
Yes | ✅ | ✅ |
| HuggingFace | huggingface |
Yes | ✅ | ✅ |
| Ollama | ollama |
No | ✅ | ✅ |
| Local | local |
No | feature-gated | feature-gated |
| Custom URL | https://... |
depends | ✅ | ✅ |
Most providers speak the OpenAI-compatible POST /embeddings shape;
HuggingFace has its own (POST /pipeline/feature-extraction/{model})
and RedDB ships a dedicated client for it. Anthropic does not have an
embeddings API — RedDB rejects embedding calls against it explicitly
rather than silently re-routing to a different provider. local
requires the local-models feature flag at engine build time.
See docs/guides/ai-providers.md
for the routing matrix, the wire shape per provider, and the
Anthropic-embeddings policy in detail.
ASK 'summarize alerts' USING groq MODEL 'llama-3.3-70b-versatile'
ASK 'summarize alerts' USING ollama MODEL 'llama3'
ASK 'summarize alerts' USING anthropicSet a default provider so you can drop USING from every query:
# Set default provider -- no more USING on every query
curl -X POST http://127.0.0.1:8080/ai/credentials \
-d '{"provider":"groq","api_key":"gsk_xxx","default":true}'-- Now ASK uses groq by default
ASK 'what happened?'# Export/import all config as JSON
curl http://127.0.0.1:8080/configBuilt-in approximate data structures for real-time analytics at scale.
-- HyperLogLog: count unique visitors (~0.8% error, ~16KB memory)
CREATE HLL visitors
HLL ADD visitors 'user1' 'user2' 'user3'
HLL COUNT visitors
-- Count-Min Sketch: frequency estimation
CREATE SKETCH click_counter WIDTH 2000 DEPTH 7
SKETCH ADD click_counter 'button_a' 5
SKETCH COUNT click_counter 'button_a'
-- Cuckoo Filter: membership testing with deletion (unlike Bloom filters)
CREATE FILTER active_sessions CAPACITY 500000
FILTER ADD active_sessions 'session_abc'
FILTER CHECK active_sessions 'session_abc'
FILTER DELETE active_sessions 'session_abc'Beyond B-tree. Create the right index for your workload.
-- Hash index: O(1) exact-match lookups
CREATE INDEX idx_email ON users (email) USING HASH
-- Bitmap index: fast analytical queries on low-cardinality columns
CREATE INDEX idx_status ON orders (status) USING BITMAP
-- R-Tree: spatial queries on geo data
CREATE INDEX idx_loc ON sites (location) USING RTREE
SEARCH SPATIAL RADIUS 48.8566 2.3522 10.0 COLLECTION sites COLUMN location LIMIT 50
SEARCH SPATIAL NEAREST 48.8566 2.3522 K 5 COLLECTION sites COLUMN locationRedDB extends SQL with WITH clauses for operational semantics:
-- TTL: auto-expire records
INSERT INTO sessions (token) VALUES ('abc') WITH TTL 1 h
-- Context indexes for cross-model search
CREATE TABLE customers (passport TEXT) WITH CONTEXT INDEX ON (passport)
-- Graph expansion inline with SELECT
SELECT * FROM users WITH EXPAND GRAPH DEPTH 2
-- Metadata on write
INSERT INTO logs (msg) VALUES ('deploy') WITH METADATA (source = 'ci')
-- Absolute expiration
INSERT INTO events (name) VALUES ('launch') WITH EXPIRES AT 1735689600000Write in whatever you think in. The engine auto-detects the language.
| Language | Example |
|---|---|
| SQL | SELECT * FROM hosts WHERE os = 'linux' |
| Cypher | MATCH (a:User)-[:FOLLOWS]->(b) RETURN b.name |
| Gremlin | g.V().hasLabel('person').out('FOLLOWS').values('name') |
| SPARQL | SELECT ?name WHERE { ?p :name ?name } |
| Natural Language | show me all critical hosts |
| ASK (RAG) | ASK 'what changed in the last 24 hours?' |
All six hit the same engine, same data, same indexes.
Stop reaching for Flyway, Liquibase, Drizzle Migrate, or Sequelize migrations. RedDB handles schema and data migrations as first-class SQL commands.
-- Register a migration
CREATE MIGRATION add_users_table AS
CREATE TABLE users (id BIGINT, email TEXT, created_at TIMESTAMP);
-- Register a dependent migration (RedDB also auto-infers deps from SQL body)
CREATE MIGRATION add_users_index DEPENDS ON add_users_table AS
CREATE INDEX idx_email ON users (email);
-- Apply everything in dependency order
APPLY MIGRATION *
-- Large data backfill in safe 5,000-row batches — resumes on crash
CREATE MIGRATION backfill_display_names BATCH 5000 ROWS AS
UPDATE users SET display_name = email WHERE display_name IS NULL;
-- Undo an applied migration (VCS revert under the hood)
ROLLBACK MIGRATION add_users_index
-- Inspect what a migration will do
EXPLAIN MIGRATION backfill_display_namesEvery applied migration creates a VCS commit (RedDB's "Git for Data"). Rollback reverts that commit automatically — no rollback scripts to maintain. Dependency ordering is a DAG; RedDB detects cycles at CREATE time and auto-infers edges from your SQL body so you rarely need explicit DEPENDS ON.
Not just TEXT and INTEGER. RedDB understands your domain.
Network: IpAddr, Ipv4, Ipv6, MacAddr, Cidr, Subnet, Port
Geo: Latitude, Longitude, GeoPoint
Locale: Country2, Country3, Lang2, Lang5, Currency
Identity: Uuid, Email, Url, Phone, Semver
Visual: Color, ColorAlpha
Cross-model refs: NodeRef, EdgeRef, VectorRef, RowRef, KeyRef, DocRef, TableRef, PageRef
Primitives: Integer, UnsignedInteger, Float, Decimal, BigInt, Text, Blob, Boolean, Json, Array, Enum
Temporal: Timestamp, TimestampMs, Date, Time, Duration
Validation on write. No parsing in your app.
Built-in backup scheduler, WAL archiving, Change Data Capture (CDC), and Point-in-Time Recovery framework:
# Poll real-time changes
curl 'localhost:8080/changes?since_lsn=0'
# Trigger manual backup
curl -X POST localhost:8080/backup/trigger
# Check backup status
curl localhost:8080/backup/statusRemote backends: S3, R2, DigitalOcean Spaces, GCS, Turso, Cloudflare D1, local filesystem.
For concrete RTO/RPO numbers per failure mode (process crash, disk loss,
PITR rollback, replica promotion), see
docs/operations/rto-rpo.md.
Every collection doubles as a key-value store with dedicated REST endpoints:
# Write a key
curl -X PUT http://127.0.0.1:8080/collections/settings/kvs/theme \
-H 'content-type: application/json' -d '{"value": "dark"}'
# Read a key
curl http://127.0.0.1:8080/collections/settings/kvs/theme
# Delete a key
curl -X DELETE http://127.0.0.1:8080/collections/settings/kvs/themeConfig keys work the same way -- read, write, or delete any red_config setting at runtime:
# Set a config key
curl -X PUT http://127.0.0.1:8080/config/red.ai.default.provider \
-d '{"value": "groq"}'
# Read a config key
curl http://127.0.0.1:8080/config/red.ai.default.provider
# Or manage config from SQL
SET CONFIG red.ai.default.provider = 'groq'
SHOW CONFIG red.ai| Mode | Think of it as... | Access via |
|---|---|---|
| Embedded | SQLite | Rust API -- RedDB::open("data.rdb") |
| Server | Postgres | HTTP + gRPC -- dual-stack |
| Agent | MCP tool | red mcp -- AI agent integration |
Same storage format across all three. Start embedded, scale to server, expose to agents -- no migration.
Where RedDB wins. Two scenarios in the canonical
duel-officialbenchmark show measurable wins over Postgres and Mongo today:
typed_insert— RedDB ≈ 16× faster than PostgreSQL on typed single-row inserts.disk_usage— RedDB ≈ 1.5× faster than MongoDB on the compact-write path.See
docs/perf/wins.mdfor the cited sessions and reproducible commands. The honest counterpart — gaps where RedDB is still behind — lives atdocs/perf/when-not-reddb.md.
RedDB uses multiple optimization techniques for fast queries at scale:
- Result Cache -- identical SELECT queries return in <1ms; auto-invalidated on INSERT/UPDATE/DELETE (30s TTL, max 1000 entries)
- Hot Item Cache --
get_any(rid)lookups served from an LRU cache (10K entries), O(1) instead of scanning all collections - Binary Bulk Insert -- gRPC
BulkInsertBinarywith zero JSON overhead, protobuf native types -- 241K ops/sec - Concurrent HTTP -- thread-per-connection model; each request handled in its own OS thread
- Parallel Segment Scanning -- sealed segments scanned in parallel via
std::thread::scope; auto-detects single-core and skips parallelism - Hash Join -- O(n+m) joins instead of O(n*m), auto-selected for large datasets
- Lazy Graph Materialization -- only loads reachable nodes instead of full graph
- Pre-filtered Vector Search -- metadata filters applied before HNSW indexing
- Index-Assisted Scans -- bloom filter + hash index hints for WHERE clauses
- Column Projection Pushdown -- only materializes SELECT columns
- Query Plan Caching -- LRU cache with 1h TTL for repeated queries
- Batch Entity Lookup -- multi-entity fetches resolved in a single pass
- Background Maintenance Thread -- backup scheduling, retention, and checkpoint run off the hot path
RedDB uses 7 layers of protection to keep your data safe:
| Layer | What it does |
|---|---|
| File Lock | Exclusive flock prevents two processes from writing the same .rdb file |
| Double-Write Buffer | Pages written to .rdb-dwb first; survives torn writes on power loss |
| Header Shadow | Copy of page 0 in .rdb-hdr; auto-recovers if header corrupts |
| Metadata Shadow | Copy of page 1 in .rdb-meta; auto-recovers collection registry |
| fsync Discipline | All critical writes followed by sync_all() (not just flush) |
| Two-Phase Checkpoint | Crash-safe WAL→DB transfer with checkpoint_in_progress flag |
| Binary Store CRC32 | V3 files have CRC32 footer + atomic write-to-temp-then-rename |
Every page has a CRC32 checksum (verified on read). Every WAL record has a CRC32 checksum. The binary store format (V3) includes a full-file CRC32 footer.
RedDB supports per-field eventual consistency via an append-only transaction log with periodic consolidation. Inspired by CRDT principles (commutative, associative reducers), it enables high-throughput write patterns while guaranteeing convergence.
# Track clicks with async consolidation (returns instantly)
curl -X POST localhost:8080/ec/urls/clicks/add -d '{"id": 1, "value": 1}'
# Check consolidated + pending value
curl localhost:8080/ec/urls/clicks/status?id=1| Feature | Description |
|---|---|
| 6 reducers | Sum, Max, Min, Count, Average, Last (last-write-wins) |
| Sync mode | Consolidates immediately (strong consistency) |
| Async mode | Background worker consolidates periodically (high throughput) |
| Transaction log | Immutable append-only audit trail per field |
| SET checkpoint | Resets base value, discards prior operations |
| All modes | Works in server, embedded (Rust API), and serverless |
See the Eventual Consistency Guide for the theory (CAP theorem, CRDTs, convergence) and full API reference.
Built-in geo functions with no external dependencies. Supports both spherical (Haversine) and ellipsoidal (Vincenty/WGS-84) models.
-- Distance from each store to a point (in km)
SELECT name, GEO_DISTANCE(location, POINT(-23.55, -46.63)) AS dist
FROM stores ORDER BY dist
-- Vincenty for sub-millimeter accuracy
SELECT name, GEO_DISTANCE_VINCENTY(location, POINT(40.71, -74.00)) AS dist
FROM airports# HTTP API
curl -X POST localhost:8080/geo/distance -d '{
"from": {"lat": -23.55, "lon": -46.63},
"to": {"lat": -22.91, "lon": -43.17}
}'| Function | What it computes |
|---|---|
GEO_DISTANCE |
Haversine distance (km) |
GEO_DISTANCE_VINCENTY |
WGS-84 geodesic distance (km) |
GEO_BEARING |
Compass direction (degrees) |
GEO_MIDPOINT |
Great-circle midpoint |
Also available: destination point, bounding box, polygon area, spatial search (RADIUS, BBOX, NEAREST). See the Geo Operations Guide.
Standalone K-Means and DBSCAN clustering on vector collections, with SIMD-accelerated distance computation and automatic parallelization.
# K-Means: group products into 5 clusters
curl -X POST localhost:8080/vectors/cluster -d '{
"collection": "products", "algorithm": "kmeans", "k": 5
}'
# DBSCAN: discover clusters automatically (no K needed)
curl -X POST localhost:8080/vectors/cluster -d '{
"collection": "products", "algorithm": "dbscan", "eps": 0.5, "min_points": 3
}'K-Means uses parallel assignment (multi-threaded for datasets > 1K vectors). DBSCAN labels unreachable points as noise (-1), useful for outlier detection. See the Vector Clustering Guide.
One connection-string API, four languages. Every driver accepts the same
connect(uri) contract so application code ports across runtimes with zero
ceremony.
| Language | Package | Install | Backends |
|---|---|---|---|
| Rust | reddb-io-client |
cargo add reddb-io-client |
embedded ✅ · gRPC ✅ · HTTP ✅ |
| Node / Bun / Deno | @reddb-io/sdk (npm) |
pnpm add @reddb-io/sdk |
stdio subprocess ✅ |
| Python | reddb (PyPI) |
pip install reddb (soon) |
embedded ✅ · gRPC ✅ · wire ✅ |
All drivers accept the same URIs:
memory:// ephemeral in-memory
file:///absolute/path embedded engine on disk
grpc://host:port remote server (planned — tracked in PLAN_DRIVERS.md)
Example — the same app in three languages:
// Rust
let db = reddb_client::Reddb::connect("memory://").await?;
db.insert("users", &JsonValue::object([("name", JsonValue::string("Alice"))])).await?;
let rows = db.query("SELECT * FROM users").await?;// Node, Bun, Deno
import { connect } from '@reddb-io/sdk'
const db = await connect('memory://')
await db.insert('users', { name: 'Alice' })
const rows = await db.query('SELECT * FROM users')# Python
import reddb
with reddb.connect("memory://") as db:
db.insert("users", {"name": "Alice"})
print(db.query("SELECT * FROM users"))Driver docs live in crates/reddb-client/README.md, drivers/js/README.md, and
drivers/python/README.md. The full protocol spec and roadmap are in
PLAN_DRIVERS.md.
For JavaScript and TypeScript, RedDB ships three packages under the
@reddb-io/ scope. Pick the one that matches your scenario — see the
JavaScript / TypeScript driver guide
for the full matrix and ADR 0007
for the rationale.
# App code in Node, Bun, or Deno — full SDK with embedded, gRPC, and HTTP transports
pnpm add @reddb-io/sdk
# Thin remote-only client for serverless, edge, CI, or sidecar runtimes (~5 MB)
pnpm add @reddb-io/client
# CLI launcher — installs the `red` binary on PATH
pnpm add -g @reddb-io/cliApplication code with the SDK:
import { connect } from '@reddb-io/sdk'
const db = await connect('memory://')
const result = await db.query('SELECT * FROM users')
await db.close()Launch the server from npm without a separate install step:
npx @reddb-io/cli@latest version
npx @reddb-io/cli@latest server --wire-bind 127.0.0.1:5050 --http-bind 127.0.0.1:8080 --path ./data.rdb# Install
curl -fsSL https://raw.githubusercontent.com/reddb-io/reddb/main/install.sh | bash
# Start the server (wire: 5050, gRPC: 5055, HTTP: 8080)
red server --wire-bind 127.0.0.1:5050 --grpc-bind 127.0.0.1:5055 --http-bind 127.0.0.1:8080 --path ./data.rdb
# Insert data
curl -X POST http://127.0.0.1:8080/query \
-H 'content-type: application/json' \
-d '{"query":"INSERT INTO hosts (ip, os) VALUES ('\''10.0.0.1'\'', '\''linux'\'')"}'
# Query it
curl -X POST http://127.0.0.1:8080/query \
-H 'content-type: application/json' \
-d '{"query":"SELECT * FROM hosts"}'Or via npm CLI launcher:
npx @reddb-io/cli@latest server --wire-bind 127.0.0.1:5050 --http-bind 127.0.0.1:8080Or via Docker:
echo "$GITHUB_TOKEN" | docker login ghcr.io -u "$GITHUB_USER" --password-stdin # if GHCR requires auth
docker run --rm -p 5050:5050 -p 5055:5055 -p 8080:8080 ghcr.io/reddb-io/reddb:latestOr, if you only need the thin remote-only client (~7 MB image):
echo "$GITHUB_TOKEN" | docker login ghcr.io -u "$GITHUB_USER" --password-stdin # if GHCR requires auth
docker run --rm ghcr.io/reddb-io/reddb-client:latest red://reddb.example.com:5050 -c "SELECT 1"For production-secure Docker (vault + secrets) and Kubernetes, see
docs/getting-started/docker.md
and docs/security/vault.md.
RedDB ships as a Cargo workspace. The reddb crate is the
umbrella that hosts the red binary; the engine, thin client,
gRPC stubs, and wire vocabulary live in sibling crates. See the
workspace migration guide
for what moved where and which crate to pick when depending on
RedDB from another Rust project.
- Documentation
- GitHub
- npm driver package
- npm CLI launcher
- Releases
- Workspace migration guide
- Public Contract Matrix — every doc'd promise mapped to a test or a planned-status note
AGPL-3.0 License -- Built by RedDB.io