AI agents that get smarter with every task ๐ง
Agentic Context Engine learns from your agent's successes and failures. Just plug in and watch your agents improve.
Star โญ๏ธ this repo if you find it useful!
- Direct your favorite coding agent (Cursor, Claude Code, Codex, etc) to Agents.md
- Prompt away!
pip install ace-frameworkexport OPENAI_API_KEY="your-api-key"
# Or use Claude, Gemini, or 100+ other providersfrom ace import LiteLLMClient, Generator, Reflector, Curator, Playbook
# Initialize with any LLM
client = LiteLLMClient(model="gpt-4o-mini")
generator = Generator(client)
reflector = Reflector(client)
curator = Curator(client)
playbook = Playbook()
# Teach your agent through examples
# (See examples/ folder for complete training patterns)
# Now it can solve new problems with learned strategies
result = generator.generate(
    question="Give me the seahorse emoji",
    context="",
    playbook=playbook
)
print(result.final_answer)  # Agent applies learned strategiesThat's it! Your agent is now learning and improving. ๐
AI agents make the same mistakes repeatedly.
ACE enables agents to learn from execution feedback: what works, what doesn't, and continuously improve. 
 No training data, no fine-tuning, just automatic improvement.
- ๐ 20-35% Better Performance: Proven improvements on complex tasks
- ๐ง Self-Improving: Agents get smarter with each task
- ๐ No Context Collapse: Preserves valuable knowledge over time
- ๐ 100+ LLM Providers: Works with OpenAI, Anthropic, Google, and more
- ๐ Production Observability: Built-in Opik integration for enterprise monitoring
A challenge where LLMs often hallucinate that a seahorse emoji exists (it doesn't). Watch ACE learn from its own mistakes in real-time. This demo shows how ACE handles the infamous challenge!
In this example:
- Round 1: The agent incorrectly outputs ๐ด (horse emoji)
- Self-Reflection: ACE reflects without any external feedback
- Round 2: With learned strategies from ACE, the agent successfully realizes there is no seahorse emoji
Try it yourself:
python examples/kayba_ace_test.pyBased on the ACE research framework from Stanford & SambaNova.
ACE uses three specialized roles that work together:
- ๐ฏ Generator - Executes tasks using learned strategies from the playbook
- ๐ Reflector - Analyzes what worked and what didn't after each execution
- ๐ Curator - Updates the playbook with new strategies based on reflection
ACE teaches your agent and internalises:
- โ Successes โ Extract patterns that work
- โ Failures โ Learn what to avoid
- ๐ง Tool usage โ Discover which tools work best for which tasks
- ๐ฏ Edge cases โ Remember rare scenarios and how to handle them
The magic happens in the Playbookโa living document of strategies that evolves with experience. 
Key innovation: All learning happens in context through incremental updatesโno fine-tuning, no training data, and complete transparency into what your agent learned.
---
config:
  look: neo
  theme: neutral
---
flowchart LR
    Playbook[("`**๐ Playbook**<br>(Evolving Context)<br><br>โขStrategy Bullets<br> โ Helpful strategies <br>โ Harmful patterns <br>โ Neutral observations`")]
    Start(["**๐Query** <br>User prompt or question"]) --> Generator["**โ๏ธGenerator** <br>Executes task using playbook"]
    Generator --> Reflector
    Playbook -. Provides Context .-> Generator
    Environment["**๐ Task Environment**<br>Evaluates answer<br>Provides feedback"] -- Feedback+ <br>Optional Ground Truth --> Reflector
    Reflector["**๐ Reflector**<br>Analyzes and provides feedback what was helpful/harmful"]
    Reflector --> Curator["**๐ Curator**<br>Produces improvement deltas"]
    Curator --> DeltaOps["**๐Merger** <br>Updates the playbook with deltas"]
    DeltaOps -- Incremental<br>Updates --> Playbook
    Generator <--> Environment
    # Basic installation
pip install ace-framework
# With demo support (browser automation)
pip install ace-framework[demos]
# With LangChain support
pip install ace-framework[langchain]
# With local model support
pip install ace-framework[transformers]
# With all features
pip install ace-framework[all]
# Development
pip install ace-framework[dev]
# Development from source (contributors) - UV Method (10-100x faster)
git clone https://github.com/kayba-ai/agentic-context-engine
cd agentic-context-engine
uv sync
# Development from source (contributors) - Traditional Method
git clone https://github.com/kayba-ai/agentic-context-engine
cd agentic-context-engine
pip install -e .ACE works with any LLM provider through LiteLLM:
# OpenAI
client = LiteLLMClient(model="gpt-4o")
# Anthropic Claude
client = LiteLLMClient(model="claude-3-5-sonnet-20241022")
# Google Gemini
client = LiteLLMClient(model="gemini-pro")
# Ollama (local)
client = LiteLLMClient(model="ollama/llama2")
# With fallbacks for reliability
client = LiteLLMClient(
    model="gpt-4",
    fallbacks=["claude-3-haiku", "gpt-3.5-turbo"]
)ACE includes built-in Opik integration for production monitoring and debugging.
# Install with Opik support
pip install ace-framework opik
# Set your Opik API key (or use local deployment)
export OPIK_API_KEY="your-api-key"
export OPIK_PROJECT_NAME="ace-project"When Opik is available, ACE automatically logs:
- Generator: Input questions, reasoning, and final answers
- Reflector: Error analysis and bullet classifications
- Curator: Playbook updates and delta operations
- Playbook Evolution: Changes to strategies over time
# Opik tracing is automatic - just run your ACE code normally
from ace import Generator, Reflector, Curator, Playbook
from ace.llm_providers import LiteLLMClient
# All role interactions are automatically tracked
generator = Generator(llm_client)
output = generator.generate(
    question="What is 2+2?",
    context="Show your work",
    playbook=playbook
)
# View traces at https://www.comet.com/opik or your local Opik instanceIf Opik is not installed or configured, ACE continues to work normally without tracing. No code changes needed.
- Quick Start Guide - Get running in 5 minutes
- API Reference - Complete API documentation
- Examples - Ready-to-run code examples
- Prompt Engineering - Advanced prompt techniques
- Changelog - See recent changes
We love contributions! Check out our Contributing Guide to get started.
Based on the ACE paper and inspired by Dynamic Cheatsheet.
If you use ACE in your research, please cite:
@article{zhang2024ace,title={Agentic Context Engineering},author={Zhang et al.},journal={arXiv:2510.04618},year={2024}}โญ Star this repo if you find it useful! 
Built with โค๏ธ by Kayba and the open-source community.

