Note
The MCP server is currently available under https://gemini-mcp-server-231532712093.europe-west1.run.app/mcp/
. It is deployed to Google Cloud Run and can be authenticated using an AI Studio API key. see examples/test_remote.py for an example on how to use the server with the google-genai
client.
A Model Context Protocol server that provides access to Google's Gemini API. This server enables LLMs to perform intelligent web searches, generate content, and access other Gemini features. It supports both STDIO and streamable-http transport modes and can be run locally or remotely. If you use STDIO mode it will try to use the GEMINI_API_KEY
environment variable. If you use streamable-http mode it will try to use the Bearer token in the Authorization header.
Available Tools:
- web_search - Performs a web search using the configurable Gemini web search model and returns synthesized results with citations
query
(string, required): The search query to executeinclude_citations
(boolean, optional): Whether to include citations in the response. Default isFalse
.
- use_gemini - Delegates a task to a specified Gemini model (Pro or Flash).
prompt
(string, required): The prompt or task for Gemini.model
(string, optional): The Gemini model to use. Uses the configured default model if not specified.
pip install git+https://github.com/philbug/gemini-mcp-server.git
- STDIO mode: Uses
GEMINI_API_KEY
environment variable - HTTP mode: Requires Bearer token in Authorization header
The Gemini MCP Server supports configurable models for different use cases. You can customize which Gemini models are used for specific tasks by setting environment variables.
The server supports three configurable model types:
- Web Search Model: Used for web search functionality
- Default Model: Used for general tasks when no specific model is requested
- Advanced Model: Used for complex tasks requiring advanced reasoning
Variable | Default Value | Description |
---|---|---|
GEMINI_WEB_SEARCH_MODEL |
gemini-flash-lite-latest |
Model to use for web search functionality |
GEMINI_DEFAULT_MODEL |
gemini-flash-latest |
Default model for general use |
GEMINI_ADVANCED_MODEL |
gemini-2.5-pro |
Advanced model for complex tasks |
Claude Code example:
"gemini-search": {
"command": "uv",
"args": [
"--directory",
"/home/{your cloned repo path}",
"run",
"python3",
"-m",
"gemini_mcp.server",
"--transport",
"stdio"
],
"env": {
"GEMINI_API_KEY": "your_api_key",
"GEMINI_WEB_SEARCH_MODEL": "gemini-flash-latest",
"GEMINI_DEFAULT_MODEL": "gemini-flash-lite-latest",
"GEMINI_ADVANCED_MODEL": "gemini-flash-latest"
}
},
The server includes improved handling of citation metadata. It now properly processes cases where grounding metadata might be None, ensuring more reliable citation generation and preventing errors during web search operations.
The server maintains full backward compatibility with existing configurations. If you don't set these environment variables, the server will use the default values shown above.
# Basic usage with default models
GEMINI_API_KEY="your_gemini_api_key_here" gemini-mcp --transport stdio
gemini-mcp --transport streamable-http
The server will start on http://0.0.0.0:8000/mcp/
You can deploy the Gemini MCP Server as Remote MCP Server to Google Cloud Run to make it available easily available to any client.
To deploy the server, run the following command from your terminal, replacing [PROJECT-ID]
and [REGION]
with your Google Cloud project ID and desired region:
# Set your project ID and region
export PROJECT_ID=remote-mcp-test-462811
export REGION=europe-west1
export SERVICE_NAME=gemini-mcp-server
# Authenticate with Google Cloud
gcloud auth login
gcloud config set project $PROJECT_ID
# Enable required services
gcloud services enable run.googleapis.com artifactregistry.googleapis.com cloudbuild.googleapis.com
# Deploy the service
gcloud run deploy $SERVICE_NAME \
--source . \
--region $REGION \
--port 8000 \
--allow-unauthenticated
The command will build the Docker image, push it to Google Artifact Registry, and deploy it to Cloud Run. After the deployment is complete, you will get a URL for your service. We will allow unauthenticated access to the service this means that anyone with the URL can send requests to the server, which it self is protected by an Authorization header. If you want to secure the service you can follow the instructions in the Cloud Run documentation.
SERVICE_NAME=gemini-mcp-server
REGION=europe-west1
gcloud run services delete $SERVICE_NAME --region $REGION
Add to your mcpServers
configuration:
STDIO Mode:
{
"mcpServers": {
"gemini-search": {
"command": "gemini-mcp",
"args": ["--transport", "stdio"],
"env": {
"GEMINI_API_KEY": "your_gemini_api_key_here",
"GEMINI_WEB_SEARCH_MODEL": "gemini-flash-latest",
"GEMINI_DEFAULT_MODEL": "gemini-flash-lite-latest",
"GEMINI_ADVANCED_MODEL": "gemini-2.5-pro"
}
}
}
}
HTTP Mode:
{
"mcpServers": {
"gemini-mcp": {
"url": "https://remote-mcp-test.com/mcp/", // replace with your remote mcp server url
"headers": { "Authorization": "Bearer YOUR_KEY" } // replace with your AI Studio API key
}
}
}
or check out the example in the examples/test_remote.py file.
from mcp.client.streamable_http import streamablehttp_client
remote_url = "https://remote-mcp-test.com/mcp/" # replace with your remote mcp server url
async with streamablehttp_client(
remote_url, headers={"Authorization": f"Bearer {api_key}"}
) as (read, write, _):
Start the server with streamable-http and test your server using the MCP inspector. Alternatively start inspector and run the server with stdio.
npx @modelcontextprotocol/inspector
With include_citations
set to False
:
{
"text": "Recent advancements in AI include breakthrough developments in large language models, computer vision, and autonomous systems..."
}
With include_citations
set to True
:
{
"text": "Recent advancements in AI include breakthrough developments in large language models, computer vision, and autonomous systems...",
"web_search_queries": ["latest AI developments 2024", "AI breakthroughs"],
"citations": [
{
"start_index": 24,
"end_index": 56,
"sources": [
{
"title": "Latest AI Developments 2024",
"uri": "https://example.com/ai-news"
}
...
],
"text": "breakthrough developments in large language models"
},
...
]
}
{
"text": "The capital of France is Paris."
}
To run the tests, run the following command from the root directory:
Note: You need to set the GEMINI_API_KEY
environment variable to run the tests.
pytest
This project is licensed under the MIT License.