PrimisAI Nexus is a powerful and flexible Python package for managing AI agents and coordinating complex tasks using LLMs. It provides a robust framework for creating, managing, and interacting with multiple specialized AI agents under the supervision of a central coordinator.
nexus-demo.mp4
- AI Base Class: A foundational class for AI interactions.
- Agent Class: Extends the AI base class with additional features for specialized tasks.
- Supervisor Class: Manages multiple agents, coordinates tasks, and handles user interactions.
- Hierarchical Supervision: Support for main and assistant supervisors enabling complex task hierarchies.
- Persistent History: Built-in conversation history management with JSONL storage.
- Integrated Logging: Organized logging system within workflow structure.
- Debugger Utility: Integrated debugging capabilities for logging and troubleshooting.
- Structured Agent Outputs: Support for schema-defined, structured responses with validation.
- Flexible Configuration: Easy-to-use configuration options for language models and agents.
- Flexible LLM Parameters: Direct control over all language model parameters through configuration.
- Interactive Sessions: Built-in support for interactive chat sessions with the AI system.
- YAML Configuration: Define complex agent hierarchies using YAML files for easy setup and modification.
- Model Context Protocol (MCP) Integration: Support for automatic discovery and usage of external tool servers via MCP, including SSE (HTTP) and stdio (local subprocess) transports.
You can install PrimisAI Nexus directly from PyPI using pip:
pip install primisai
If you prefer to build the package from source, clone the repository and install it with pip:
git clone git@github.com:PrimisAI/nexus.git
cd nexus
pip install -e .
Here's a simple example to get you started with Nexus:
from primisai.nexus.core import AI, Agent, Supervisor
from primisai.nexus.utils.debugger import Debugger
# Configure your OpenAI API key
llm_config = {
"api_key": "your-api-key-here",
"model": "gpt-4o",
"base_url": "https://api.openai.com/v1",
}
# Create a supervisor
supervisor = Supervisor("MainSupervisor", llm_config)
# Create and register agents
agent1 = Agent("Agent1", llm_config, system_message="You are a helpful assistant.")
agent2 = Agent("Agent2", llm_config, system_message="You are a creative writer.")
supervisor.register_agent(agent1)
supervisor.register_agent(agent2)
# Start an interactive session
supervisor.display_agent_graph()
supervisor.start_interactive_session()
PrimisAI Nexus supports defining complex agent hierarchies using YAML configuration files. This feature allows for easy setup and modification of agent structures without changing the Python code.
Here's a simple example of a YAML configuration file:
supervisor:
name: TaskManager
type: supervisor
llm_config:
model: ${LLM_MODEL}
api_key: ${LLM_API_KEY}
base_url: ${LLM_BASE_URL}
system_message: "You are the task management supervisor."
children:
- name: TaskCreator
type: agent
llm_config:
model: ${LLM_MODEL}
api_key: ${LLM_API_KEY}
base_url: ${LLM_BASE_URL}
system_message: "You are responsible for creating new tasks."
keep_history: true
tools:
- name: add_task
type: function
python_path: examples.task_management_with_yaml.task_tools.add_task
The keep_history
parameter allows you to control whether an agent maintains conversation history between interactions. When set to False
, the agent treats each query independently, useful for stateless operations. When True
(default), the agent maintains context from previous interactions.
To use this YAML configuration:
from primisai.nexus.config import load_yaml_config, AgentFactory
# Load the YAML configuration
config = load_yaml_config('path/to/your/config.yaml')
# Create the agent structure
factory = AgentFactory()
task_manager = factory.create_from_config(config)
# Start an interactive session
task_manager.start_interactive_session()
For a more detailed example of YAML configuration, check out the task management example.
- Flexibility: Easily modify agent structures without changing Python code.
- Readability: YAML configurations are human-readable and easy to understand.
- Scalability: Define complex hierarchies of supervisors and agents in a clear, structured manner.
- Separation of Concerns: Keep agent definitions separate from application logic.
For detailed documentation on each module and class, please refer to the inline docstrings in the source code.
PrimisAI Nexus provides comprehensive history management and logging capabilities organized within workflow directories:
nexus_workflows/
├── workflow_123/ # Workflow specific directory
│ ├── history.jsonl # Conversation history
│ └── logs/ # Workflow logs
│ ├── MainSupervisor.log
│ ├── AssistantSupervisor.log
│ └── Agent1.log
└── standalone_logs/ # Logs for agents not in workflows
└── StandaloneAgent.log
You can restore any agent or supervisor's LLM-compatible context with a single call, enabling true warm starts and reproducibility, even for multi-level workflows.
from primisai.nexus.history import HistoryManager
manager = HistoryManager(workflow_id)
supervisor.chat_history = manager.load_chat_history("SupervisorName")
agent.chat_history = manager.load_chat_history("AgentName")
This ensures that only the relevant delegated turns, tool calls, and responses are loaded for each entity, preserving correct and replayable LLM state across runs.
PrimisAI Nexus allows for complex interactions between multiple agents. You can create specialized agents for different tasks, register them with a supervisor, and let the supervisor manage the flow of information and task delegation.
# Example of creating a specialized agent with tools
tools = [
{
"metadata": {
"name": "search_tool",
"description": "Searches the internet for information"
},
"tool": some_search_function
}
]
research_agent = Agent("Researcher", llm_config, tools=tools, system_message="You are a research assistant.", use_tools=True)
supervisor.register_agent(research_agent)
PrimisAI Nexus allows agents to provide schema-validated, structured outputs. This ensures consistent response formats and enables reliable downstream processing.
# Define an output schema for a code-writing agent
code_schema = {
"type": "object",
"properties": {
"description": {
"type": "string",
"description": "Explanation of the code's purpose"
},
"code": {
"type": "string",
"description": "The actual code implementation"
},
"language": {
"type": "string",
"description": "Programming language used"
}
},
"required": ["description", "code"]
}
# Create an agent with structured output
code_agent = Agent(
name="CodeWriter",
llm_config=llm_config,
system_message="You are a skilled programmer.",
output_schema=code_schema,
strict=True # Enforce schema validation
)
# Agent responses will be automatically formatted and validated
response = code_agent.chat("Write a function to calculate factorial")
# Response will be JSON-structured:
# {
# "description": "Function to calculate factorial of a number",
# "code": "def factorial(n):\n if n <= 1: return 1\n return n * factorial(n-1)",
# "language": "python"
# }
The output_schema
parameter defines the expected structure of the agent's responses, while the strict
parameter controls validation:
- When
strict=True
, responses are guaranteed to match the schema - When
strict=False
, the agent attempts to follow the schema but falls back to unstructured responses if needed
This feature is particularly useful for:
- Ensuring consistent output formats
- Building reliable agent pipelines
- Automated processing of agent responses
- Integration with downstream systems
For detailed examples of schema usage, including complex workflows with multiple schema-aware agents, see the output schema examples and schema-aware workflow example.
PrimisAI Nexus supports a hierarchical supervisor structure with two types of supervisors:
-
Main Supervisor: The root supervisor that manages the overall workflow
-
Assistant Supervisor: Specialized supervisors that handle specific task domains
Here's how to create and use different types of supervisors:
# Create a main supervisor with a specific workflow ID
main_supervisor = Supervisor(
name="MainSupervisor",
llm_config=llm_config,
workflow_id="custom_workflow_123"
)
# Create an assistant supervisor
assistant_supervisor = Supervisor(
name="AnalysisManager",
llm_config=llm_config,
is_assistant=True,
system_message="You manage data analysis tasks."
)
# Create agents
data_agent = Agent("DataAnalyst", llm_config, system_message="You analyze data.")
viz_agent = Agent("Visualizer", llm_config, system_message="You create visualizations.")
# Register assistant supervisor with main supervisor
main_supervisor.register_agent(assistant_supervisor)
# Register agents with assistant supervisor
assistant_supervisor.register_agent(data_agent)
assistant_supervisor.register_agent(viz_agent)
# Display the complete hierarchy
main_supervisor.display_agent_graph()
The above code creates a hierarchical structure where:
- Main Supervisor manages the overall workflow
- Assistant Supervisor handles specialized tasks
- Agents perform specific operations
The display_agent_graph()
output will show:
Main Supervisor: MainSupervisor
│
└── Assistant Supervisor: AnalysisManager
│
├── Agent: DataAnalyst
│ └── No tools available
│
└── Agent: Visualizer
└── No tools available
Each workflow is automatically assigned a unique ID and maintains its conversation history in a dedicated directory structure:
custom_workflow_123/
├── history.jsonl
└── logs
├── AnalysisManager.log
├── DataAnalyst.log
├── MainSupervisor.log
└── Visualizer.log
All interactions, delegations, and tool usage are automatically logged and stored in the workflow directory, providing complete visibility into the decision-making process and execution flow.
PrimisAI Nexus supports automatic tool discovery and usage via external Model Context Protocol (MCP) servers. This enables seamless integration of local or remote tool infrastructures, including both SSE (HTTP) and stdio (local subprocess) transports.
-
SSE (HTTP/Server-Sent Events):
- Connect to any MCP-compatible HTTP server exposing tools via an SSE endpoint.
- Recommended for remote, network-accessible, or containerized tool servers.
-
Stdio (Local Subprocess):
- Launches a local MCP server using Python as a subprocess and communicates over stdin/stdout.
- Recommended for securely sandboxed or development tool servers.
When creating an Agent
, use the mcp_servers
argument to specify a list of tool servers and their transport types:
mcp_servers = [
{
"type": "sse",
"url": "http://localhost:8000/sse"
},
{
"type": "sse",
"url": "https://remote.mcpservers.org/fetch"
},
{
"type": "stdio",
"script_path": "examples/agent_with_mcp_stdio/weather_server.py"
}
]
- For
"type": "sse"
, the"url"
field must be the exact SSE endpoint provided by your MCP server (for example:/sse
,/fetch
, or another custom path). No path rewriting or appending is performed by the framework. - For
"type": "stdio"
, provide the path to your MCP server Python script as"script_path"
.
from primisai.nexus.core import Agent
agent = Agent(
name="RemoteMCPAgent",
llm_config=llm_config,
mcp_servers=[
{"type": "sse", "url": "https://remote.mcpservers.org/fetch"},
{"type": "stdio", "script_path": "weather_server.py"}
],
use_tools=True
)
agent.chat("What is 2 plus 3?")
If tools are added, removed, or modified on any MCP server during runtime, call:
agent.update_mcp_tools()
This will refresh the list of available MCP tools without restarting the agent.
- If your SSE MCP server requires authentication, add an
"auth_token"
field to the server dictionary. - You can mix and match any number of SSE and stdio MCP servers per agent.
- Tool schemas are converted automatically for use with function-calling models.
For a complete working demonstration of using a Supervisor with multiple agents, each utilizing different MCP transport mechanisms (SSE and stdio), see the example and detailed instructions in: Multiple Agents with MCP
If you find Nexus useful, please consider citing our preprint.
@article{sami2025nexus,
title={Nexus: A Lightweight and Scalable Multi-Agent Framework for Complex Tasks Automation},
author={Sami, Humza and ul Islam, Mubashir and Charas, Samy and Gandhi, Asav and Gaillardon, Pierre-Emmanuel and Tenace, Valerio},
journal={arXiv preprint arXiv:2502.19091},
year={2025}
}
This project is licensed under the MIT License.