Skip to content

my-ai-stack/stackme

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

6 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

license apache-2.0
library_name stackme
tags
memory
context
llm
rag
agentic
local-ai
open-source

🧠 Stackme

The context layer for every AI.

Stackme is a free, open-source memory layer for AI. It stores what matters about you, retrieves relevant context before every query, and injects it into any AI — ChatGPT, Claude, Copilot, Gemini, Ollama, or any AI via API.

No server. No subscription. No data leaves your machine.


Install

pip install stackme

Or install from source:

pip install git+https://github.com/my-ai-stack/stackme

Quick Start

from stackme import Context

ctx = Context()

# Add facts about yourself
ctx.add_fact("I run a fintech B2B SaaS, launched March 2024")
ctx.add_fact("Q3 goal: 10K paying customers")
ctx.add_fact("Users are 25-40, income $50-100K")

# Add a user message — facts are auto-extracted
ctx.add_user_message("I'm building a B2B SaaS, targeting fintech companies")

# Ask any AI — Stackme retrieves your context first
context = ctx.get_relevant("What pricing should we use?")
# → "## Facts\n- I run a fintech B2B SaaS...\n- Q3 goal: 10K paying customers..."

# Your AI gets the full picture every time.
print(context)

How It Works

You: "What pricing should we use?"

Stackme retrieves:
  - I run a fintech B2B SaaS
  - Q3 goal: 10K paying customers
  - Users are 25-40, income $50-100K

Enriched prompt → ChatGPT / Claude / any AI

Architecture

~/.stackme/
├── memory.sqlite     ← all memories
├── vectors.faiss     ← semantic index
└── facts.graph       ← structured knowledge graph
  • Session Memory — current conversation, in-process
  • Short-Term Memory — last 24h, SQLite
  • Long-Term Memory — permanent, SQLite + vector search
  • Knowledge Graph — structured facts, extracted from your prompts

CLI

# Add facts
stackme add-fact "I run a fintech startup"
stackme add "I'm building a B2B SaaS"   # also auto-extracts facts

# Retrieve context
stackme get "what pricing?"

# Search all memories
stackme search "fintech"

# List all facts
stackme facts

# Knowledge graph
stackme graph

# Session history
stackme history --last 10

# Export all data
stackme export

# Stats
stackme count

API Reference

Context

Main class. All data stored locally in ~/.stackme/.

ctx = Context(user_id="default")  # multi-user support

Adding Memories

ctx.add_fact("I run a fintech startup")       # structured fact
ctx.add_prompt("User asked about pricing")    # user prompt
ctx.add_context("AI responded with...")      # AI response
ctx.add_user_message("I'm building a B2B SaaS")  # both + auto-extract
ctx.add_ai_message("Here's my recommendation...")

Retrieving

ctx.get_relevant("pricing strategy?", top_k=5)  # → formatted string
ctx.search("fintech", top_k=10)                  # → list of strings
ctx.get_facts()                                   # → all facts
ctx.get_graph(subject="User")                     # → GraphFact list

Session

ctx.get_session_history(last_n=20)  # in-memory conversation
ctx.clear_session()                # wipe session only

Utilities

ctx.export()      # full JSON backup
ctx.count()       # total items
ctx.clear_all()   # ⚠️ wipe everything

Privacy

Everything stays on your machine. Your memories are yours. We never see, store, or transmit your data. No account required.


License

Apache 2.0 — free for commercial and personal use.


Built by Stack AI · GitHub · HuggingFace

About

No description, website, or topics provided.

Resources

License

Security policy

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors