LLMterface is under active development. APIs may change and the README may lag slightly behind implementation. Until v1.0, minor releases may introduce breaking changes.
A small, opinionated, provider-agnostic interface for working with LLMs in Python.
LLMterface focuses on one thing: submitting prompts to AI providers and getting back structured, validated responses.
No agents. No chains. No workflows. No hidden control flow.
If you want a thin abstraction layer you can understand in one sitting, this is for you.
LLMterface exists to solve a narrow problem: sending prompts to LLMs and getting validated responses while avoiding vendor lock-in and minimizing complexity.
It intentionally avoids orchestration, memory, agents, and workflow abstractions so those concerns remain explicit and application-owned. This makes LLMterface easy to reason about, debug, and integrate alongside other tools rather than replacing them.
- Minimal: A small, readable interface with very few moving parts.
- Generic: Universal configuration primitives that map cleanly to provider defaults, with optional provider-specific overrides.
- Unobtrusive: Designed to coexist with other LLM libraries without forcing architectural decisions.
- Extensible: Providers can be added via entry points without modifying core code.
If you want any of the following, build them on top of LLMterface or use a different library:
- Agent orchestration
- Tool calling frameworks
- Prompt chains or workflow systems
- Long-term memory systems
pip install llmterfacepip install llmterface[gemini]
pip install llmterface[openai] # TODO
pip install llmterface[openai,anthropic] # TODO
# or to install all currently supported providers
pip install llmterface[all]import llmterface as llm
handler = llm.LLMterface(
config=llm.GenericConfig(
api_key="<YOUR API KEY>",
model=llm.GenericModelType.text_lite,
provider="gemini",
)
)
question = llm.Question(
question="how many LLMs does it take to screw in a lightbulb?"
"\nExplain your reasoning.",
)
res = handler.ask(question)
print(res)
# -> “Depends. One to do it, five to argue about alignment, and twelve to hallucinate that the room is already bright.”Configuration can be supplied at three levels:
- Handler
- Chat
- Question
Overrides apply in that order, with the most specific configuration winning.
Q = "which provider are you, 'anthropic', 'openai', or 'gemini'?"
handler = llm.LLMterface(
config=llm.GenericConfig(
provider="gemini", api_key="<YOUR GEMINI API KEY>"
)
)
chat_config = llm.GenericConfig(provider="openai", api_key="<YOUR OPENAI API KEY>")
chat_id = handler.create_chat(chat_config.provider, config=chat_config)
question = llm.Question(
question=Q,
config=llm.GenericConfig(
provider="anthropic", api_key="<YOUR ANTHROPIC API KEY>"
),
)
assert handler.ask(Q).response == "gemini"
assert handler.ask(Q, chat_id=chat_id).response == "openai"
assert handler.ask(question, chat_id=chat_id).response == "anthropic"If you need access to vendor-specific features, you can supply provider overrides explicitly.
import llmterface as llm
import llmterface_gemini as gemini
gemini_override = gemini.GeminiConfig(
api_key="<YOUR GEMINI API KEY>",
model=gemini.GeminiTextModelType.CHAT_2_0_FLASH_LITE,
)
config = llm.GenericConfig(
provider=gemini.GeminiConfig.PROVIDER,
provider_overrides={
gemini.GeminiConfig.PROVIDER: gemini_override
},
)
handler = llm.LLMterface(config=config)
res = handler.ask("How many LLMs does it take to screw in a lightbulb?")
print(res.response)
# -> 6LLMterface is designed to work naturally with Pydantic models.
from pydantic import BaseModel, Field
import llmterface as llm
class WeatherResponse(BaseModel):
temperature_c: float = Field(..., description="Temperature in Celsius")
condition: str = Field(
..., description="Weather described in a silly way"
)
question = llm.Question(
question="What is the current weather in Paris?",
config=llm.GenericConfig(
api_key="<YOUR GEMINI API KEY>",
provider="gemini",
response_model=WeatherResponse,
),
)
res = llm.LLMterface().ask(question)
assert isinstance(res, WeatherResponse)
print(res.temperature_c)
# -> 12.0
print(res.condition)
# -> 'Sunny with a chance of croissants'LLMterface is built around a small set of core objects.
A Question represents a single prompt submission to an LLM, along with optional configuration, retry behavior, and response typing.
At its simplest, a Question is just text:
import llmterface as llm
question = llm.Question(
question="What is the answer to life, the universe, and everything?"
)But Question is also generic over the expected response type:
import llmterface as llm
question = llm.Question[llm.simple_answers.SimpleInteger](
question="What is 6 * 7?",
config=llm.GenericConfig(
provider="gemini",
api_key="<YOUR GEMINI API KEY>",
response_model=llm.simple_answers.SimpleInteger)
)This allows LLMterface to validate and return structured responses automatically.
Before submission, the question text is normalized via get_question():
- Dedented
- Stripped of leading/trailing whitespace
This makes multiline prompts predictable and easy to format.
You can override this behavior by subclassing Question.
Question defines how retries are handled, not the client.
Retries occur when:
-
The provider errors
-
The response fails schema validation
The on_retry() method can be overridden to implement custom retry behavior, including modifying the prompt, incorporating the previous response, or stopping retries entirely.
By default, schema validation failures cause the prompt to be augmented with a strict formatting reminder and the previous response content.
GenericConfig is the provider-agnostic configuration model used throughout LLMterface.
It defines a common set of fields that map cleanly onto most LLM providers. The goal is that your application code can stay stable even if you switch providers, because you configure intent (model tier, temperature, response model), not vendor-specific knobs.
If a field is not supported by a given provider, it is simply ignored by that provider integration.
If vendor-specific configuration is required, it can be supplied via provider_overrides.
import llmterface as llm
import llmterface_gemini as gemini
config = llm.GenericConfig(
provider="gemini",
api_key="<YOUR API KEY>",
model=llm.GenericModelType.text_lite,
temperature=0.2,
response_model=llm.simple_answers.SimpleString,
provider_overrides={
"gemini": gemini.GeminiConfig(
api_key="<YOUR GEMINI API KEY>",
# provider-specific fields here
)
},
)Configurations in LLMterface are not merged field-by-field.
When one configuration takes precedence over another, it fully replaces the lower-precedence configuration at that level.
For example:
- A question-level config completely overrides a chat-level config
- A chat-level config completely overrides a handler-level config
This keeps configuration resolution simple, explicit, and predictable.
Vendor-specific configuration objects inherit from:
llmterface.providers.provider_config.ProviderConfigProvider configs may expose additional fields beyond GenericConfig. These are only interpreted by the corresponding provider integration and are ignored elsewhere.
This design allows provider integrations to evolve independently without leaking provider-specific concerns into application code.