Skip to content

THU-MAIC/OpenMAIC

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

56 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

OpenMAIC Banner

Get an immersive, multi-agent learning experience in just one click

Paper License: AGPL-3.0 Live Demo Deploy with Vercel OpenClaw Integration Stars
Discord   Feishu
Next.js React TypeScript LangGraph Tailwind CSS

English | 简体中文
Live Demo · Quick Start · Features · Use Cases · OpenClaw

📖 Overview

OpenMAIC (Open Multi-Agent Interactive Classroom) is an open-source AI platform that turns any topic or document into a rich, interactive classroom experience. Powered by multi-agent orchestration, it generates slides, quizzes, interactive simulations, and project-based learning activities — all delivered by AI teachers and AI classmates who can speak, draw on a whiteboard, and engage in real-time discussions with you. With built-in OpenClaw integration, you can generate classrooms directly from messaging apps like Feishu, Slack, or Telegram.

promo-en.mp4

Highlights

  • One-click lesson generation — Describe a topic or attach your materials; the AI builds a full lesson in minutes
  • Multi-agent classroom — AI teachers and peers lecture, discuss, and interact with you in real time
  • Rich scene types — Slides, quizzes, interactive HTML simulations, and project-based learning (PBL)
  • Whiteboard & TTS — Agents draw diagrams, write formulas, and explain out loud
  • Export anywhere — Download editable .pptx slides or interactive .html pages
  • OpenClaw integration — Generate classrooms from Feishu, Slack, Telegram, and 20+ messaging apps via your AI assistant

Tip

Use OpenMAIC from your chat app — zero setup

With OpenClaw, you can generate classrooms directly from Feishu, Slack, Discord, Telegram, and 20+ messaging apps. Two ways to get started:

Option A: Hosted mode (no local setup needed)

  1. Get an access code at open.maic.chat
  2. clawhub install openmaic
  3. Tell your assistant "teach me quantum physics" — done!

Option B: Self-hosted — the skill walks you through clone, config, and startup step by step.

Learn more →


🚀 Quick Start

Prerequisites

  • Node.js >= 18
  • pnpm >= 10

1. Clone & Install

git clone https://github.com/THU-MAIC/OpenMAIC.git
cd OpenMAIC
pnpm install

2. Configure

cp .env.example .env.local

Fill in at least one LLM provider key:

OPENAI_API_KEY=sk-...
ANTHROPIC_API_KEY=sk-ant-...
GOOGLE_API_KEY=...

You can also configure providers via server-providers.yml:

providers:
  openai:
    apiKey: sk-...
  anthropic:
    apiKey: sk-ant-...

Supported providers: OpenAI, Anthropic, Google Gemini, DeepSeek, and any OpenAI-compatible API.

Recommended model: Gemini 3 Flash — best balance of quality and speed. For highest quality (at slower speed), try Gemini 3.1 Pro.

If you want OpenMAIC server APIs to use Gemini by default, also set DEFAULT_MODEL=google:gemini-3-flash-preview.

3. Run

pnpm dev

Open http://localhost:3000 and start learning!

4. Build for Production

pnpm build && pnpm start

Vercel Deployment

Deploy with Vercel

Or manually:

  1. Fork this repository
  2. Import into Vercel
  3. Set environment variables (at minimum one LLM API key)
  4. Deploy

Docker Deployment

cp .env.example .env.local
# Edit .env.local with your API keys, then:
docker compose up --build

Optional: MinerU (Advanced Document Parsing)

MinerU provides enhanced parsing for complex tables, formulas, and OCR. You can use the MinerU official API or self-host your own instance.

Set PDF_MINERU_BASE_URL (and PDF_MINERU_API_KEY if needed) in .env.local.


✨ Features

Lesson Generation

Describe what you want to learn or attach reference materials. OpenMAIC's two-stage pipeline handles the rest:

Stage What Happens
Outline AI analyzes your input and generates a structured lesson outline
Scenes Each outline item becomes a rich scene — slides, quizzes, interactive modules, or PBL activities

Classroom Components

🎓 Slides

AI teachers deliver lectures with voice narration, spotlight effects, and laser pointer animations — just like a real classroom.

🧪 Quiz

Interactive quizzes (single / multiple choice, short answer) with real-time AI grading and feedback.

🔬 Interactive Simulation

HTML-based interactive experiments for visual, hands-on learning — physics simulators, flowcharts, and more.

🏗️ Project-Based Learning (PBL)

Choose a role and collaborate with AI agents on structured projects with milestones and deliverables.

Multi-Agent Interaction

  • Classroom Discussion — Agents proactively initiate discussions; you can jump in anytime or get called on
  • Roundtable Debate — Multiple agents with different personas discuss a topic, with whiteboard illustrations
  • Q&A Mode — Ask questions freely; the AI teacher responds with slides, diagrams, or whiteboard drawings
  • Whiteboard — AI agents draw on a shared whiteboard in real time — solving equations step by step, sketching flowcharts, or illustrating concepts visually.

OpenClaw Integration

OpenMAIC integrates with OpenClaw — a personal AI assistant that connects to messaging platforms you already use (Feishu, Slack, Discord, Telegram, WhatsApp, etc.). With this integration, you can generate and view interactive classrooms directly from your chat app without ever touching a terminal.

Just tell your OpenClaw assistant what you want to learn — it handles everything else:

  • Hosted mode — Grab an access code from open.maic.chat, save it in your config, and generate classrooms instantly — no local setup required
  • Self-hosted mode — Clone, install dependencies, configure API keys, and start the server — the skill guides you through each step
  • Track progress — Poll the async generation job and send you the link when ready

Every step asks for your confirmation first. No black-box automation.

Available on ClawHub — Install with one command:

clawhub install openmaic

Or copy manually:

mkdir -p ~/.openclaw/skills
cp -R /path/to/OpenMAIC/skills/openmaic ~/.openclaw/skills/openmaic
Configuration & details
Phase What the skill does
Clone Detect an existing checkout or ask before cloning/installing
Startup Choose between pnpm dev, pnpm build && pnpm start, or Docker
Provider Keys Recommend a provider path; you edit .env.local yourself
Generation Submit an async generation job and poll until it completes

Optional config in ~/.openclaw/openclaw.json:

{
  "skills": {
    "entries": {
      "openmaic": {
        "config": {
          // Hosted mode: paste your access code from open.maic.chat
          "accessCode": "sk-xxx",
          // Self-hosted mode: local repo path and URL
          "repoDir": "/path/to/OpenMAIC",
          "url": "http://localhost:3000"
        }
      }
    }
  }
}

Export

Format Description
PowerPoint (.pptx) Fully editable slides with images, charts, and LaTeX formulas
Interactive HTML Self-contained web pages with interactive simulations

And More

  • Text-to-Speech — Multiple voice providers with customizable voices
  • Speech Recognition — Talk to your AI teacher using your microphone
  • Web Search — Agents search the web for up-to-date information during class
  • i18n — Interface supports Chinese and English
  • Dark Mode — Easy on the eyes for late-night study sessions

💡 Use Cases

"Teach me Python from scratch in 30 min"

"How to play the board game Avalon"

"Analyze the stock prices of Zhipu and MiniMax"

"Break down the latest DeepSeek paper"


🤝 Contributing

We welcome contributions from the community! Whether it's bug reports, feature ideas, or pull requests — every bit helps.

Project Structure

OpenMAIC/
├── app/                        # Next.js App Router
│   ├── api/                    #   Server API routes (~18 endpoints)
│   │   ├── generate/           #     Scene generation pipeline (outlines, content, images, TTS …)
│   │   ├── generate-classroom/ #     Async classroom job submission + polling
│   │   ├── chat/               #     Multi-agent discussion (SSE streaming)
│   │   ├── pbl/                #     Project-Based Learning endpoints
│   │   └── ...                 #     quiz-grade, parse-pdf, web-search, transcription, etc.
│   ├── classroom/[id]/         #   Classroom playback page
│   └── page.tsx                #   Home page (generation input)
│
├── lib/                        # Core business logic
│   ├── generation/             #   Two-stage lesson generation pipeline
│   ├── orchestration/          #   LangGraph multi-agent orchestration (director graph)
│   ├── playback/               #   Playback state machine (idle → playing → live)
│   ├── action/                 #   Action execution engine (speech, whiteboard, effects)
│   ├── ai/                     #   LLM provider abstraction
│   ├── api/                    #   Stage API facade (slide/canvas/scene manipulation)
│   ├── store/                  #   Zustand state stores
│   ├── types/                  #   Centralized TypeScript type definitions
│   ├── audio/                  #   TTS & ASR providers
│   ├── media/                  #   Image & video generation providers
│   ├── export/                 #   PPTX & HTML export
│   ├── hooks/                  #   React custom hooks (55+)
│   ├── i18n/                   #   Internationalization (zh-CN, en-US)
│   └── ...                     #   prosemirror, storage, pdf, web-search, utils
│
├── components/                 # React UI components
│   ├── slide-renderer/         #   Canvas-based slide editor & renderer
│   │   ├── Editor/Canvas/      #     Interactive editing canvas
│   │   └── components/element/ #     Element renderers (text, image, shape, table, chart …)
│   ├── scene-renderers/        #   Quiz, Interactive, PBL scene renderers
│   ├── generation/             #   Lesson generation toolbar & progress
│   ├── chat/                   #   Chat area & session management
│   ├── settings/               #   Settings panel (providers, TTS, ASR, media …)
│   ├── whiteboard/             #   SVG-based whiteboard drawing
│   ├── agent/                  #   Agent avatar, config, info bar
│   ├── ui/                     #   Base UI primitives (shadcn/ui + Radix)
│   └── ...                     #   audio, roundtable, stage, ai-elements
│
├── packages/                   # Workspace packages
│   ├── pptxgenjs/              #   Customized PowerPoint generation
│   └── mathml2omml/            #   MathML → Office Math conversion
│
├── skills/                     # OpenClaw / ClawHub skills
│   └── openmaic/               #   Guided OpenMAIC setup & generation SOP
│       ├── SKILL.md            #   Thin router with confirmation rules
│       └── references/         #   On-demand SOP sections
│
├── configs/                    # Shared constants (shapes, fonts, hotkeys, themes …)
└── public/                     # Static assets (logos, avatars)

Key Architecture

  • Generation Pipeline (lib/generation/) — Two-stage: outline generation → scene content generation
  • Multi-Agent Orchestration (lib/orchestration/) — LangGraph state machine managing agent turns and discussions
  • Playback Engine (lib/playback/) — State machine driving classroom playback and live interaction
  • Action Engine (lib/action/) — Executes 28+ action types (speech, whiteboard draw/text/shape/chart, spotlight, laser …)

How to Contribute

  1. Fork the repository
  2. Create your feature branch (git checkout -b feature/amazing-feature)
  3. Commit your changes (git commit -m 'Add amazing feature')
  4. Push to the branch (git push origin feature/amazing-feature)
  5. Open a Pull Request

💼 Commercial Licensing

This project is licensed under AGPL-3.0. For commercial licensing inquiries, please contact: thu_maic@tsinghua.edu.cn


📝 Citation

If you find OpenMAIC useful in your research, please consider citing:

@Article{JCST-2509-16000,
  title = {From MOOC to MAIC: Reimagine Online Teaching and Learning through LLM-driven Agents},
  journal = {Journal of Computer Science and Technology},
  volume = {},
  number = {},
  pages = {},
  year = {2026},
  issn = {1000-9000(Print) /1860-4749(Online)},
  doi = {10.1007/s11390-025-6000-0},
  url = {https://jcst.ict.ac.cn/en/article/doi/10.1007/s11390-025-6000-0},
  author = {Ji-Fan Yu and Daniel Zhang-Li and Zhe-Yuan Zhang and Yu-Cheng Wang and Hao-Xuan Li and Joy Jia Yin Lim and Zhan-Xin Hao and Shang-Qing Tu and Lu Zhang and Xu-Sheng Dai and Jian-Xiao Jiang and Shen Yang and Fei Qin and Ze-Kun Li and Xin Cong and Bin Xu and Lei Hou and Man-Li Li and Juan-Zi Li and Hui-Qin Liu and Yu Zhang and Zhi-Yuan Liu and Mao-Song Sun}
}

⭐ Star History

Star History Chart


📄 License

This project is licensed under the GNU Affero General Public License v3.0.

About

Open Multi-Agent Interactive Classroom — Get an immersive, multi-agent learning experience in just one click

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors