
MemFuse Python SDK
The official Python client for MemFuse, the open-source memory layer for LLMs.
Explore the Docs »
View Demo
·
Report Bug
·
Request Feature
Table of Contents
Large language model applications are inherently stateless by design. When the context window reaches its limit, previous conversations, user preferences, and critical information simply disappear.
MemFuse bridges this gap by providing a persistent, queryable memory layer between your LLM and storage backend, enabling AI agents to:
- Remember user preferences and context across sessions
- Recall facts and events from thousands of interactions later
- Optimize token usage by avoiding redundant chat history resending
- Learn continuously and improve performance over time
This repository contains the official Python SDK for seamless integration with MemFuse servers. For comprehensive information about the MemFuse server architecture and advanced features, please visit the MemFuse Server repository.
Note: This is the standalone Client SDK repository. If you need to install and run the MemFuse server, which is essential to use the SDK, please visit the MemFuse Server repository.
You can install the MemFuse Python SDK using one of the following methods:
Option 1: Install from PyPI (Recommended)
pip install memfuse
Option 2: Install from Source
git clone https://github.com/memfuse/memfuse-python.git
cd memfuse-python
pip install -e .
Here's a comprehensive example demonstrating how to use the MemFuse Python SDK with OpenAI:
from memfuse.llm import OpenAI
from memfuse import MemFuse
import os
memfuse_client = MemFuse(
# api_key=os.getenv("MEMFUSE_API_KEY")
# base_url=os.getenv("MEMFUSE_BASE_URL"),
)
memory = memfuse_client.init(
user="alice",
# agent="agent_default",
# session=<randomly-generated-uuid>
)
# Initialize your LLM client with the memory scope
llm_client = OpenAI(
api_key=os.getenv("OPENAI_API_KEY"), # Your OpenAI API key
memory=memory
)
# Make a chat completion request
response = llm_client.chat.completions.create(
model="gpt-4o", # Or any model supported by your LLM provider
messages=[{"role": "user", "content": "I'm planning a trip to Mars. What is the gravity there?"}]
)
print(f"Response: {response.choices[0].message.content}")
# Example Output: Response: Mars has a gravity of about 3.721 m/s², which is about 38% of Earth's gravity.
Now, ask a follow-up question. MemFuse will automatically recall relevant context from the previous conversation:
# Ask a follow-up question. MemFuse automatically recalls relevant context.
followup_response = llm_client.chat.completions.create(
model="gpt-4o",
messages=[{"role": "user", "content": "What are some challenges of living on that planet?"}]
)
print(f"Follow-up: {followup_response.choices[0].message.content}")
# Example Output: Follow-up: Some challenges of living on Mars include its thin atmosphere, extreme temperatures, high radiation levels, and the lack of liquid water on the surface.
MemFuse automatically manages the retrieval of relevant information and storage of new memories from conversations within the specified memory
scope.
Explore comprehensive examples in the examples/ directory of this repository, featuring:
- Basic Operations: Fundamental usage patterns and asynchronous operations
- Conversation Continuity: Maintaining context across multiple interactions
- UI Integrations: Gradio-based chatbot implementations with streaming support
- Server Documentation: For detailed information about the MemFuse server architecture and advanced configuration, visit the MemFuse online documentation
- SDK Documentation: Comprehensive API references and guides will be available soon
Join our growing community:
- GitHub Discussions: Participate in roadmap discussions, RFCs, and Q&A in the MemFuse Server repository
- Issues & Features: Report bugs or request features in this repository's Python SDK Issues section
If MemFuse enhances your projects, please ⭐ star both the server repository and this SDK repository!
This MemFuse Python SDK is licensed under the Apache 2.0 License. See the LICENSE file for complete details.