Language: English | 切换到简体中文
OpenCane is an AI backend runtime for smart-cane scenarios, focused on an end-to-end loop: device connectivity -> realtime conversation -> visual lifelog -> digital tasks -> safety and observability.
- Product goal: build an OpenClaw experience for visually impaired users
- Interaction carrier: smart-cane hardware
- Human-device interaction: voice + buttons (input), voice + vibration (feedback)
- Core loop: hardware access -> realtime conversation -> visual memory -> digital tasks -> safety/observability
- Runtime form: backend service with control APIs and multi-modem adapter layer
- Multi-adapter hardware ingress:
mock / websocket / ec600 / generic_mqtt - Multi-modem support in
generic_mqtt:ec600mcnle_v1 / a7670c_v1 / sim7600g_h_v1 / ec800m_v1 / ml307r_dl_v1 - Realtime voice pipeline: segmentation, VAD, transcription, playback, and interruption
- Visual lifelog: async image ingestion, semantic retrieval, timeline retrieval, and safety annotations
- Digital task execution: create, query, cancel, offline pushback, and retry
- Control APIs: device registration/binding, operation dispatch, and runtime status queries
- On-the-go voice assistance: press-to-talk interaction for quick answers while walking.
- Forward risk check: photo + voice queries to identify nearby obstacles and hazards.
- Short-term memory recall: ask about recent moments via timeline and semantic retrieval.
- Digital task delegation: turn voice intents into trackable tasks with status updates.
- Device health and offline alerts: detect heartbeat/connection anomalies and notify users.
- Emergency help trigger (SOS): minimal-action emergency trigger with immediate confirmation.
- Digital Journal (emotion-first): multimodal models understand daily photo sequences and voice context to form long-term memory; it can be replayed with personalized, podcast-like narration and shared as text with family and friends.
- Layered architecture:
adapter / runtime / agent / api / storage / safety - Tool execution strategy: MCP-first with automatic fallback to
web_search / web_fetch / exec - Task lifecycle:
pending -> running -> success/failed/timeout/canceled - Persistent data layer: SQLite for lifelog, tasks, and observability
- Vector retrieval backends:
chromaandqdrant - Model provider abstraction via LiteLLM:
OpenAI / Anthropic / Gemini / DashScopeand more - Runtime observability with realtime metrics and historical samples
- Config governance with built-in
dev/staging/prodprofiles andconfig check --strict
git clone https://github.com/iflabx/opencane.git
cd opencane
pip install -e .First-time initialization:
opencane onboardApply a profile (staging recommended first):
opencane config profile apply --profile CONFIG_PROFILE_STAGING.json
opencane config check --strictStart hardware runtime (example):
opencane hardware serve --adapter mock --logs- Overview:
docs/overview.md - Quickstart:
docs/quickstart.md - Architecture:
docs/architecture.md - Data Flow:
docs/data-flow.md - Hardware Runtime:
docs/hardware-runtime.md - Control API:
docs/api/control.md - Lifelog API:
docs/api/lifelog.md - Digital Task API:
docs/api/digital-task.md - Deployment & Config:
docs/deployment-config.md - Operations Runbook:
docs/operations-runbook.md - Security Baseline:
docs/security.md - Roadmap:
docs/roadmap.md
- Primary CLI command:
opencane - Python package/import path:
opencane - Default data directory:
~/.opencane - Local internal docs directory:
local-docs/
OpenCane is continuously evolved from an upstream open-source foundation by HKUDS.
Thanks to HKUDS for the original engineering base.
语言:简体中文 | Switch to English
OpenCane 是一个面向智能盲杖场景的 AI 后端运行时,聚焦“设备接入 -> 实时对话 -> 图像记忆 -> 数字任务 -> 安全与观测”的完整闭环。
- 产品目标:打造“给视障人士使用的 OpenClaw”
- 交互载体:智能盲杖硬件
- 人机交互:语音 + 按钮输入,语音 + 震动反馈
- 核心闭环:设备接入 -> 实时对话 -> 图像记忆 -> 数字任务 -> 安全与观测
- 系统形态:后端运行时 + 控制 API + 多蜂窝模组适配层
- 多硬件接入:支持
mock / websocket / ec600 / generic_mqtt适配器统一接入 - 多模组适配:
generic_mqtt内置ec600mcnle_v1 / a7670c_v1 / sim7600g_h_v1 / ec800m_v1 / ml307r_dl_v1 - 实时语音链路:支持音频分段、VAD、转写、播报与打断处理
- 视觉 Lifelog:图像异步入库、语义检索、时间线检索与安全标注
- 数字任务执行:任务创建、状态查询、取消、离线回推与重试
- 控制面 API:设备注册绑定、设备指令下发、运行状态查询
- 行走中即时问答:按键触发语音提问,低延迟语音回复,不打断通行节奏。
- 前方风险确认:拍照 + 语音询问,识别附近障碍与风险并给出提示。
- 短时记忆回溯:支持“刚才发生了什么”的时间线与语义检索回答。
- 数字任务代办:将语音需求转成可追踪任务,并持续回报状态。
- 设备异常与离线提醒:心跳或连接异常时及时告警,恢复后可继续会话。
- 紧急求助触发(SOS):通过最少动作触发求助,并立即得到确认反馈。
- 数字日记(情感优先):多模态大模型理解每天的照片序列与语音形成长期记忆;可用个性化语音像播客一样回放,也可生成文本分享给亲人朋友。
- 分层架构:
adapter / runtime / agent / api / storage / safety - 工具执行策略:优先 MCP 工具,失败自动回退
web_search / web_fetch / exec - 任务状态机:
pending -> running -> success/failed/timeout/canceled - 数据存储:SQLite 持久化(lifelog / task / observability)
- 向量检索:支持
chroma与qdrant后端 - 提供商适配:基于 LiteLLM 统一接入
OpenAI / Anthropic / Gemini / DashScope等模型 - 运维可观测:运行时指标 + 历史观测数据,支持问题复盘
- 配置治理:内置
dev/staging/prod模板与config check --strict
git clone https://github.com/iflabx/opencane.git
cd opencane
pip install -e .首次初始化:
opencane onboard应用配置模板(建议先用 staging 模板):
opencane config profile apply --profile CONFIG_PROFILE_STAGING.json
opencane config check --strict启动硬件运行时(示例):
opencane hardware serve --adapter mock --logs- 总览:
docs/overview.md - 快速开始:
docs/quickstart.md - 架构设计:
docs/architecture.md - 数据流:
docs/data-flow.md - 硬件运行时:
docs/hardware-runtime.md - 控制 API:
docs/api/control.md - Lifelog API:
docs/api/lifelog.md - Digital Task API:
docs/api/digital-task.md - 部署与配置:
docs/deployment-config.md - 运维手册:
docs/operations-runbook.md - 安全基线:
docs/security.md - 路线图:
docs/roadmap.md
- CLI 主命令:
opencane - Python 包与导入路径:
opencane - 默认数据目录:
~/.opencane - 本地内部文档目录:
local-docs/
本项目基于 HKUDS 的上游开源基础进行持续开发与场景化演进。
感谢 HKUDS 团队提供的原始工程基础。
