GPU-accelerated force-directed graph | Multi-user immersive XR | 71 agent skills | OWL 2 ontology governance
Why VisionClaw? | Quick Start | Architecture | Documentation | Contributing
They just haven't told you yet.
73% of frontline AI adoption happens without management sign-off. Your workforce is already building shadow workflows — stitching together AI agents, automating procurement shortcuts, inventing cross-functional pipelines that don't appear on any org chart. The question isn't whether your organisation is becoming an agentic mesh. It's whether you'll shape how it forms — or be reshaped by it.
The personal agent revolution has a governance problem. Tools like Claude Code have shown that autonomous AI agents are powerful, popular, and ready to act. They've also shown what happens when agents operate without shared semantics, formal reasoning, or organisational guardrails: unauthorised actions, prompt injection attacks, and enterprises deploying security scanners just to detect rogue agent instances on their own networks.
VisionClaw takes the opposite approach. Governance isn't the brake — it's what lets you drive at 200 mph.
VisionClaw is an open-source platform that transforms organisations into governed agentic meshes — where autonomous AI agents, human judgment, and institutional knowledge work together through a shared semantic substrate.
VisionClaw gives organisations a governed intelligence layer where 71 specialist agent skills reason over a formal OWL 2 ontology before they act. Every agent decision is semantically grounded, every mutation passes consistency checking, and every reasoning chain is auditable from edge case back to first principles. The middle manager doesn't disappear — they evolve into the Judgment Broker, reviewing only the genuine edge cases and strategic decisions that exceed agent authority.
The platform ingests knowledge from Logseq notebooks via GitHub, reasons over it with an OWL 2 EL inference engine (Whelk), renders the result as an interactive 3D graph where nodes attract or repel based on their semantic relationships, and exposes everything to AI agents through 7 Model Context Protocol tools. Users collaborate in the same space through multi-user XR presence, spatial voice, and immersive graph exploration.
VisionClaw is built on a Rust/Actix-web backend with hexagonal architecture, a React 19 + Three.js frontend with WebGPU/WebGL dual rendering, Neo4j graph storage, and CUDA 13.1 GPU-accelerated physics. It is production-proven — currently operational at MediaCityUK, powering a 15-person creative technology team, and validated in partnership with a major UK creative studio and the University of Salford.
GPU-accelerated force-directed graph layout with real-time physics controls — 934 nodes responding to spring/repulsion forces
Interacting with a knowledge graph in an immersive projected environment
git clone https://github.com/DreamLab-AI/VisionClaw.git
cd VisionClaw && cp .env.example .env
docker-compose --profile dev up -d| Service | URL | Description |
|---|---|---|
| Frontend | http://localhost:3001 | 3D knowledge graph interface |
| API | http://localhost:4000/api | REST and WebSocket endpoints |
| Neo4j Browser | http://localhost:7474 | Graph database explorer |
| Vircadia Server | ws://localhost:3020/world/ws | Multi-user WebSocket endpoint |
Enable voice routing
docker-compose -f docker-compose.yml -f docker-compose.voice.yml --profile dev up -dAdds LiveKit SFU (port 7880), turbo-whisper STT (CUDA), and Kokoro TTS.
Enable multi-user XR
docker-compose -f docker-compose.yml -f docker-compose.vircadia.yml --profile dev up -dAdds Vircadia World Server with avatar sync, spatial audio, and collaborative graph editing.
Native build (Rust + CUDA)
curl --proto '=https' --tlsv1.2 -sSf https://sh.rustup.rs | sh
git clone https://github.com/DreamLab-AI/VisionClaw.git
cd VisionClaw && cp .env.example .env
cargo build --release --features gpu
cd client && npm install && npm run build && cd ..
./target/release/webxrVisionClaw implements a three-layer agentic mesh architecture. Insights bubble up from frontline discovery, are orchestrated through formal semantic pipelines, and governed by declarative policy — with humans as the irreplaceable judgment layer at the top.
flowchart TB
subgraph Layer3["LAYER 3 — DECLARATIVE GOVERNANCE"]
JB["Judgment Broker\n(Human-in-the-Loop)"]
Policy["AI-Enforced Policies\nBias · Security · Alignment"]
Trust["Cascading Trust\nDID/Nostr Identity"]
end
subgraph Layer2["LAYER 2 — ORCHESTRATION"]
Skills["71 Agent Skills\nClaude-Flow DAG Pipelines"]
Ontology["OWL 2 EL Reasoning\nWhelk Inference Engine"]
MCP["7 MCP Tools\nKnowledge Graph Read/Write"]
GPU["GPU Compute\nCUDA 13.1 Kernels"]
end
subgraph Layer1["LAYER 1 — DISCOVERY ENGINE"]
Ingest["Knowledge Ingestion\nLogseq · GitHub · RSS"]
Graph["Neo4j Knowledge Graph\n+ Qdrant Vectors"]
Viz["3D Visualisation\nReact 19 · Three.js · WebXR"]
Voice["Voice Routing\n4-Plane Architecture"]
end
Layer1 -->|"Insights bubble up"| Layer2
Layer2 -->|"Exceptions surface"| Layer3
Layer3 -->|"Governance flows down"| Layer2
Layer2 -->|"Validated workflows deploy"| Layer1
style Layer3 fill:#1A0A2A,stroke:#8B5CF6
style Layer2 fill:#0A1A2A,stroke:#00D4FF
style Layer1 fill:#0A2A1A,stroke:#10B981
The discovery layer ingests, structures, and renders organisational knowledge as a navigable, interactive 3D space.
Ontology Pipeline — VisionClaw syncs Logseq markdown from GitHub, parses OWL 2 EL axioms, runs Whelk inference for subsumption and consistency checking, and stores results in both Neo4j (persistent) and an in-memory OntologyRepository (fast access). GPU semantic forces use the ontology to drive graph layout physics — subClassOf creates attraction, disjointWith creates repulsion.
flowchart LR
GH["GitHub\n(Logseq)"] -->|sync| Parser["OWL Parser\n(assembler → converter)"]
Parser --> Whelk["Whelk Reasoner\n(EL++ inference)"]
Whelk --> Store["OntologyRepository\n(In-Memory)"]
Store --> Neo4j[(Neo4j)]
Store --> Physics["Semantic Forces\n(GPU)"]
Store --> Agents["Agent Tools\n(MCP)"]
Explore a live ontology dataset at narrativegoldmine.com — a 2D interactive graph and data explorer built on the same ontology data that VisionClaw renders in 3D.
3D Knowledge Visualisation — Dual-renderer architecture: WebGPU with Three Shading Language (TSL) materials on supported browsers, automatic WebGL fallback on others. Runtime toggle in the Effects panel. Post-processing via GemPostProcessing with bloom, colour grading, and depth effects.
Multi-User XR — Vircadia World Server provides spatial presence, collaborative graph editing, and avatar synchronisation. The client detects Quest headsets and applies XR optimisations: foveated rendering, DPR capping, dynamic resolution scaling. CollaborativeGraphSync handles multi-user selections, annotations, and conflict resolution. The immersive graph technology was validated in the THG world record attempt — a large-scale multi-user data visualisation event demonstrating real-time collaborative graph exploration at scale.
Rendering materials
| Material | Effect |
|---|---|
GemNodeMaterial |
Primary node material with analytics-driven colour |
CrystalOrbMaterial |
Depth-pulsing emissive with cosmic spectrum + Fresnel |
AgentCapsuleMaterial |
Bioluminescent heartbeat pulse driven by activity level |
GlassEdgeMaterial |
Animated flow emissive for relationship edges |
Voice routing (4-plane architecture)
| Plane | Direction | Scope | Trigger |
|---|---|---|---|
| 1 | User mic → turbo-whisper STT → Agent | Private | PTT held |
| 2 | Agent → Kokoro TTS → User ear | Private | Agent responds |
| 3 | User mic → LiveKit SFU → All users | Public (spatial) | PTT released |
| 4 | Agent TTS → LiveKit → All users | Public (spatial) | Agent configured public |
Opus 48kHz mono end-to-end. HRTF spatial panning from Vircadia entity positions.
Logseq ontology input (source data)
| Ontology metadata | Graph structure |
|---|---|
![]() |
![]() |
| OWL entity page with category, hierarchy, and source metadata | Graph view showing semantic clusters |
Dense knowledge graph in Logseq — the raw ontology that VisionClaw ingests, reasons over, and renders in 3D
The orchestration layer is where agents reason, coordinate, and act — always against the shared semantic substrate of the OWL 2 ontology.
71 Specialist Agent Skills — The multi-agent-docker/ container provides a complete AI orchestration environment with Claude-Flow coordination and 71 skill modules spanning creative production, research, knowledge codification, governance, workflow discovery, financial intelligence, spatial/immersive, and identity/trust domains.
Why OWL 2 Is the Secret Weapon — Most agentic systems fail at scale because they lack a shared language. In VisionClaw, agents don't "guess" what a concept means — they reason against a common OWL 2 ontology. The same concept of "deliverable" means the same thing to a Creative Production agent and a Governance agent. Agent skill routing isn't keyword matching — it's ontological subsumption. The orchestration layer knows that a "risk assessment" is a sub-task of "governance review", and routes accordingly.
7 Ontology Agent Tools (MCP) — Read/write access to the knowledge graph via Model Context Protocol:
| Tool | Purpose |
|---|---|
ontology_discover |
Semantic keyword search with Whelk inference expansion |
ontology_read |
Enriched note with axioms, relationships, schema context |
ontology_query |
Validated Cypher execution with schema-aware label checking |
ontology_traverse |
BFS graph traversal from starting IRI |
ontology_propose |
Create/amend notes → consistency check → GitHub PR |
ontology_validate |
Axiom consistency check against Whelk reasoner |
ontology_status |
Service health and statistics |
GPU-Accelerated Physics — CUDA 13.1 kernels run server-authoritative force-directed graph layout. The ForceComputeActor dispatches GPU computation with real-time parameter updates. Clients receive position updates via the binary protocol V5 and apply optimistic tweening for smooth 60 FPS rendering. The BroadcastOptimizer uses delta compression with full-broadcast triggers on parameter changes.
| Metric | Result |
|---|---|
| GPU vs CPU speedup | 55x |
| Position update size | 48 bytes/node (V3) |
| WebSocket latency | 10ms |
| Binary vs JSON bandwidth | 80% reduction |
Agent skill domains (71 skills)
Creative Production — Script, storyboard, shot-list, grade & publish workflows. ComfyUI orchestration for image, video, and 3D asset generation via containerised API middleware.
Research & Synthesis — Multi-source ingestion, GraphRAG, semantic clustering, Perplexity integration.
Knowledge Codification — Tacit-to-explicit extraction; OWL concept mapping; Logseq-formatted output.
Governance & Audit — Bias detection, provenance chains (content-addressed beads), declarative policy enforcement.
Workflow Discovery — Shadow workflow detection; DAG proposal & validation.
Financial Intelligence — R&D tax modelling, grant pipeline, ROI attribution.
Spatial & Immersive — XR scene graph, light field, WebXR rendering agent, Blender, ComfyUI 3D.
Identity & Trust — DID management, key rotation, Nostr agent communications.
Development & Quality — Rust development, pair programming, agentic QE fleet, GitHub code review, performance analysis.
Infrastructure & DevOps — Docker management, Kubernetes ops, Linux admin, network analysis, monitoring.
Document Processing — LaTeX, DOCX, XLSX, PPTX, PDF generation and manipulation.
Binary WebSocket Protocol V5
High-frequency updates use a compact binary protocol instead of JSON, achieving 80% bandwidth reduction.
| Type | Code | Size | Purpose |
|---|---|---|---|
POSITION_UPDATE |
0x10 |
48 bytes/node | Node positions from GPU physics |
AGENT_POSITIONS |
0x11 |
Variable | Batch agent position updates |
VELOCITY_UPDATE |
0x12 |
Variable | Node velocity vectors |
AGENT_STATE_FULL |
0x20 |
Variable | Complete agent state snapshot |
AGENT_STATE_DELTA |
0x21 |
Variable | Incremental agent state |
GRAPH_UPDATE |
0x01 |
Variable | Graph topology changes |
VOICE_DATA |
0x02 |
Variable | Opus audio frames |
SYNC_UPDATE |
0x50 |
Variable | Multi-user sync |
SELECTION_UPDATE |
0x52 |
Variable | Collaborative selection |
VR_PRESENCE |
0x54 |
Variable | XR avatar positions |
HEARTBEAT |
0x33 |
1 byte | Connection keepalive |
BACKPRESSURE_ACK |
0x34 |
Variable | Flow control |
Features: delta encoding, flate2 streaming compression, path-registry ID compression, node type flag bits (bits 26-31) for agent/knowledge/ontology classification.
Actor system (40+ actors)
The backend uses Actix actors for supervised concurrency. GPU compute actors run physics simulations, while service actors coordinate ontology processing, client sessions, and voice routing.
GPU Compute Actors:
| Actor | Purpose |
|---|---|
ForceComputeActor |
Core force-directed layout (CUDA) |
StressMajorizationActor |
Stress majorisation algorithm |
ClusteringActor |
Graph clustering |
PageRankActor |
PageRank computation |
ShortestPathActor |
Single-source shortest paths |
ConnectedComponentsActor |
Component detection |
AnomalyDetectionActor |
Outlier node detection |
SemanticForcesActor |
OWL-driven attraction/repulsion |
ConstraintActor |
Layout constraint solving |
AnalyticsSupervisor |
GPU analytics orchestration |
Service Actors:
| Actor | Purpose |
|---|---|
GraphStateActor |
Canonical graph state |
OntologyActor |
OWL class management |
WorkspaceActor |
Multi-workspace isolation |
ClientCoordinatorActor |
Per-client session management |
PhysicsOrchestratorActor |
GPU physics delegation |
SemanticProcessorActor |
NLP query processing |
VoiceCommandsActor |
Voice-to-action routing |
TaskOrchestratorActor |
Background task scheduling |
The governance layer is what separates VisionClaw from every "move fast and break things" agent framework. Policies are code. Bias thresholds, access controls, and audit trails are embedded into every DAG transition — not bolted on afterwards.
The Judgment Broker — The middle manager doesn't disappear in the agentic mesh. They evolve. Freed from reporting, forecasting, and coordination (all now automated), the Judgment Broker focuses on three irreplaceable human capacities:
- Strategic direction — Only humans can decide what the organisation should be doing next year, and whether the mesh is pointed at it.
- Ethical adjudication — Bias, fairness, and consequence live in human judgment. No agent should be the final word on edge cases.
- Relational intelligence — Trust, culture, and coalition-building across departments are the lubrication layer no algorithm can replicate.
HITL by Design — The Human-in-the-Loop is not a fallback. It's an architectural feature. Agents know their authority boundary and surface exceptions cleanly. Every ontology mutation passes through a GitHub pull request workflow, giving human reviewers full visibility and veto over structural changes before they are committed.
Ontological Provenance — Every agent decision traces back through the OWL 2 knowledge graph. Auditors can traverse the full reasoning chain — agent-by-agent, task-by-task. Every action is recorded as an immutable bead — a content-addressed, cryptographically verifiable unit of provenance — stored alongside human-readable Markdown summaries.
Cascading Trust Hierarchies — DID-based agent identities with W3C-compliant key rotation. When an agent is revoked, the revocation cascades through dependent agents automatically. Self-sovereign infrastructure means no single point of trust failure.
Mesh KPIs — measuring what matters in a governed agentic organisation
| KPI | Formula | Target | What It Measures |
|---|---|---|---|
| Mesh Velocity | Δt(insight → codified workflow) | < 48h | How fast a discovered shortcut becomes a sanctioned, reusable DAG. |
| Augmentation Ratio | Cognitive load offloaded ÷ Total cognitive load | > 65% | What percentage of decision-making is handled by agents without human escalation. |
| Trust Variance | σ(Agent Decision Quality) over 30-day window | < 0.12σ | Drift or bias monitoring in the automated task layer. |
| HITL Precision | Correct escalations ÷ Total escalations | > 90% | Are the edge cases the mesh flags actually requiring human intervention? |
How shadow workflows become sanctioned organisational intelligence:
┌─────────────┐ ┌─────────────────┐ ┌──────────────┐ ┌──────────────┐ ┌───────────────┐
│ DISCOVERY │────▶│ CODIFICATION │────▶│ VALIDATION │────▶│ INTEGRATION │────▶│ AMPLIFICATION │
│ │ │ │ │ │ │ │ │ │
│ Passive agent│ │ IRIS maps the │ │ The Judgment │ │ Promoted to │ │ Mesh propaga- │
│ monitoring │ │ new path as a │ │ Broker │ │ live mesh │ │ tes pattern │
│ detects the │ │ proposed DAG — │ │ reviews for │ │ with SLAs, │ │ to other │
│ pattern │ │ OWL 2 formalised│ │ strategic │ │ ownership, │ │ teams where │
│ │ │ with provenance │ │ fit & bias │ │ quality │ │ it applies │
└─────────────┘ └─────────────────┘ └──────────────┘ └──────────────┘ └───────────────┘
flowchart TB
subgraph Client["Browser Client (React 19 + Three.js)"]
R3F["React Three Fiber"]
Materials["TSL + WebGL Materials"]
BinProto["Binary Protocol V5"]
Voice["Voice Orchestrator"]
XR["WebXR / Quest 3"]
Vircadia["Vircadia Services"]
end
subgraph Server["Rust Backend (Actix-web)"]
Handlers["HTTP/WS Handlers"]
Actors["40+ Concurrent Actors"]
Services["Ontology Pipeline"]
AudioRouter["Audio Router"]
MCP["MCP Tool Server"]
end
subgraph Data["Data Layer"]
Neo4j[(Neo4j 5.13)]
PG[(PostgreSQL)]
Qdrant[(Qdrant Vectors)]
end
subgraph GPU["GPU Compute (CUDA 13.1)"]
Kernels["CUDA Kernels"]
Physics["Force Simulation"]
Analytics["Graph Analytics"]
end
subgraph Agents["Multi-Agent Stack"]
Skills["71 Agent Skills"]
ClaudeFlow["Claude-Flow Orchestrator"]
AgenticQE["Agentic QE Fleet"]
end
Client <-->|"Binary V5 + SQL-over-WS"| Server
Voice <-->|"LiveKit SFU + Opus"| AudioRouter
Vircadia <-->|"WebRTC P2P"| Vircadia
Server <--> Neo4j
Server <--> Qdrant
Server <--> GPU
MCP <--> Agents
Agents -->|"GitHub PRs"| Services
style Client fill:#e1f5ff,stroke:#0288d1
style Server fill:#fff3e0,stroke:#ff9800
style Data fill:#f3e5f5,stroke:#9c27b0
style GPU fill:#e8f5e9,stroke:#4caf50
style Agents fill:#fce4ec,stroke:#e91e63
Hexagonal architecture (Ports & Adapters)
VisionClaw follows strict hexagonal architecture. Business logic in src/services/ depends only on port traits defined in src/ports/. Concrete implementations live in src/adapters/, swapped at startup via dependency injection.
flowchart LR
subgraph Ports["src/ports/ (Traits)"]
GP[GraphRepository]
KG[KnowledgeGraphRepository]
OR[OntologyRepository]
IE[InferenceEngine]
GPA[GpuPhysicsAdapter]
GSA[GpuSemanticAnalyzer]
SR[SettingsRepository]
end
subgraph Adapters["src/adapters/ (Implementations)"]
Neo4jGraph[Neo4jGraphRepository]
Neo4jOntology[Neo4jOntologyRepository]
Whelk[WhelkInferenceEngine]
CudaPhysics[PhysicsOrchestratorAdapter]
CudaSemantic[GpuSemanticAnalyzerAdapter]
Neo4jSettings[Neo4jSettingsRepository]
end
subgraph Services["src/services/ (Business Logic)"]
OQS[OntologyQueryService]
OMS[OntologyMutationService]
GPS[GitHubPRService]
OPS[OntologyPipelineService]
end
Services --> Ports
Adapters -.->|implements| Ports
style Ports fill:#e8f5e9,stroke:#4caf50
style Adapters fill:#fff3e0,stroke:#ff9800
style Services fill:#e1f5ff,stroke:#0288d1
| Port Trait | Adapter | Purpose |
|---|---|---|
GraphRepository |
ActorGraphRepository |
Graph CRUD via actor messages |
KnowledgeGraphRepository |
Neo4jGraphRepository |
Neo4j Cypher queries |
OntologyRepository |
Neo4jOntologyRepository |
OWL class/axiom storage |
InferenceEngine |
WhelkInferenceEngine |
OWL 2 EL reasoning |
GpuPhysicsAdapter |
PhysicsOrchestratorAdapter |
CUDA force simulation |
GpuSemanticAnalyzer |
GpuSemanticAnalyzerAdapter |
GPU semantic forces |
SettingsRepository |
Neo4jSettingsRepository |
Persistent settings |
| Metric | Result | Conditions |
|---|---|---|
| GPU physics speedup | 55x | vs single-threaded CPU |
| WebSocket latency | 10ms | Local network |
| Bandwidth reduction | 80% | Binary V5 vs JSON |
| Concurrent users | 250+ | Vircadia World Server |
| Position update size | 48 bytes | Per node (V3 protocol) |
| Agent concurrency | 50+ | Via actor supervisor tree |
Full technology breakdown
| Layer | Technology | Detail |
|---|---|---|
| Backend | Rust (2021 edition), Actix-web | 427 files, 175K LOC, hexagonal architecture |
| Frontend | React 19, Three.js 0.182, R3F | 370 files, 96K LOC, TypeScript 5.9 |
| Graph DB | Neo4j 5.13 | Primary store, Cypher queries, bolt protocol |
| Relational DB | PostgreSQL 15 | Vircadia World Server entity storage |
| Vector DB | Qdrant | Semantic similarity search |
| GPU | CUDA 13.1 | GPU compute via cudarc/cust crates |
| Ontology | OWL 2 EL, Whelk-rs | EL++ subsumption, consistency checking (20 source files) |
| XR | WebXR, @react-three/xr | Meta Quest 3, hand tracking, foveated rendering |
| Multi-User | Vircadia World Server | Avatar sync, spatial audio, entity CRUD |
| Voice | LiveKit SFU | turbo-whisper STT, Kokoro TTS, Opus codec |
| Protocol | Binary V5 | 48-byte position updates, delta encoding, flag-bit node typing |
| Auth | Nostr NIP-07/NIP-98 | Browser extension signing, relay integration |
| Agents | MCP, Claude-Flow | 71 skills, 7 ontology tools |
| AI/ML | GraphRAG, RAGFlow | Knowledge retrieval, inference |
| Build | Vite 6, Vitest, Playwright | Frontend build, unit tests, E2E tests |
| Infra | Docker Compose | 10 compose files, multi-profile deployment |
| CI | GitHub Actions | Build, test, docs quality, ontology federation |
VisionClaw uses the Diataxis documentation framework — 285 markdown files organised into four categories:
| Category | Path | Content |
|---|---|---|
| Tutorials | docs/tutorials/ |
First graph, digital twin, protein folding, multiplayer |
| How-To Guides | docs/how-to/ |
Deployment, agents, features, operations, development |
| Explanation | docs/explanation/ |
Architecture, concepts, system overview, design decisions |
| Reference | docs/reference/ |
API specs, database schemas, port/adapter catalogue |
Key entry points:
- Full Documentation Hub
- Architecture Overview
- Project Structure
- Ontology Agent Tools
- Voice Routing
- Docker Compose Guide
- Hexagonal Architecture
- Rust (2021 edition) with cargo
- Node.js 20+ with npm
- Docker and Docker Compose
- CUDA 13.1 (optional, for GPU acceleration)
# Backend
cargo build --release
cargo test
# Frontend
cd client && npm install && npm run build && npm test
# Integration tests
cargo test --test ontology_agent_integration_testEnvironment variables
Copy .env.example and configure:
| Variable | Description |
|---|---|
NEO4J_URI |
Neo4j bolt connection (default: bolt://localhost:7687) |
NEO4J_USER / NEO4J_PASSWORD |
Neo4j credentials |
VITE_VIRCADIA_ENABLED |
Enable Vircadia multi-user (true/false) |
VITE_VIRCADIA_SERVER_URL |
Vircadia World Server WebSocket URL |
VITE_VIRCADIA_AUTH_TOKEN |
Vircadia authentication token |
VITE_VIRCADIA_AUTH_PROVIDER |
Auth provider (system or nostr) |
VITE_VIRCADIA_ENABLE_SPATIAL_AUDIO |
Enable HRTF spatial audio |
VITE_QUEST3_ENABLE_HAND_TRACKING |
Enable Quest 3 hand tracking |
LIVEKIT_URL |
LiveKit server URL for voice routing |
LIVEKIT_API_KEY / LIVEKIT_API_SECRET |
LiveKit credentials |
GITHUB_TOKEN |
GitHub token for ontology PR creation |
GITHUB_OWNER / GITHUB_REPO |
Target repository for ontology PRs |
| Tier | CPU | RAM | GPU | Use Case |
|---|---|---|---|---|
| Minimum | 4-core 2.5GHz | 8 GB | Integrated | Development, < 10K nodes |
| Recommended | 8-core 3.0GHz | 16 GB | GTX 1060 / RX 580 | Production, < 50K nodes |
| Enterprise | 16+ cores | 32 GB+ | RTX 4080+ (16GB VRAM) | Large graphs, multi-user XR |
Platform Support: Linux (full GPU), macOS (CPU-only), Windows (WSL2), Meta Quest 3 (Beta)
VisionClaw/
├── src/ # Rust backend (427 files, 175K LOC)
│ ├── actors/ # 40+ Actix actors (GPU compute + services)
│ ├── adapters/ # Neo4j, Whelk, CUDA adapter implementations
│ ├── handlers/ # HTTP/WebSocket request handlers
│ ├── services/ # Business logic (ontology, voice, agents)
│ ├── ports/ # Trait definitions (hexagonal boundaries)
│ ├── gpu/ # CUDA kernel bridge, memory, streaming
│ ├── ontology/ # OWL parser, reasoning, physics integration
│ ├── protocols/ # Binary settings protocol
│ ├── models/ # Data models
│ └── config/ # Configuration management
├── client/ # React frontend (370 files, 96K LOC)
│ └── src/
│ ├── features/ # 13 feature modules (graph, settings, etc.)
│ ├── services/ # Voice, WebSocket, auth, integration services
│ ├── rendering/ # Custom TSL materials, post-processing
│ └── immersive/ # XR/VR specific code
├── multi-agent-docker/ # AI agent orchestration container
│ ├── skills/ # 71 agent skill modules
│ ├── mcp-infrastructure/ # MCP servers, config, tools
│ └── management-api/ # Agent lifecycle management
├── docs/ # Diataxis documentation (285 files)
├── tests/ # Integration tests
├── config/ # LiveKit, deployment config
└── scripts/ # Build, migration, testing scripts
See the Contributing Guide for development workflow, branching conventions, and coding standards.
Mozilla Public License 2.0 — Use commercially, modify freely, share changes to MPL files.
VisionClaw is built by DreamLab AI Consulting residential laboratory.

