Home / MCP / Ollama MCP Server

Ollama MCP Server

Provides tools to manage Ollama models, run inferences, and access Ollama Cloud features from MCP clients.

typescript
Installation
Add the following to your MCP client configuration file.

Configuration

View docs
{
    "mcpServers": {
        "ollama": {
            "command": "npx",
            "args": [
                "-y",
                "ollama-mcp"
            ],
            "env": {
                "OLLAMA_HOST": "http://127.0.0.1:11434",
                "OLLAMA_API_KEY": "YOUR_API_KEY"
            }
        }
    }
}

This MCP server exposes Ollama’s SDK as MCP tools so your local LLM models can be controlled by MCP-compatible apps like Claude Desktop and Cline. It automates tool discovery, supports cloud and local models, and lets you combine fast local inference with cloud capabilities in a single, easy-to-use endpoint.

How to use

You connect from an MCP client to the Ollama MCP Server to access a full set of tools for managing models, running inferences, and performing web-enabled operations through Ollama Cloud. Use the tools to list, inspect, pull, push, copy, and delete models; run text generation, chat, and embeddings; and perform web searches and fetches when you enable Ollama Cloud access. You can mix local and cloud models in a single server to balance privacy and connectivity, and you can integrate with Claude Desktop, Cline, and other MCP clients without writing custom adapters.

How to install

Prerequisites: you need Node.js v16 or newer and a working Ollama installation (local or cloud access). You also need npm or pnpm to install dependencies and manage packages.

Install the MCP server globally so you can run it from any project location.

npm install -g ollama-mcp

Available tools

ollama_list

List all available local models in Ollama.

ollama_show

Get detailed information about a specific model.

ollama_pull

Download models from the Ollama library.

ollama_push

Push models to the Ollama library.

ollama_copy

Create a copy of an existing model.

ollama_delete

Remove models from local storage.

ollama_create

Create custom models from a Modelfile.

ollama_ps

List models currently running in Ollama.

ollama_generate

Generate text completions from a model.

ollama_chat

Interactive chat with models, with support for tools and functions.

ollama_embed

Generate embeddings for provided text.

ollama_web_search

Search the web via Ollama Cloud with configurable results (requires API key).

ollama_web_fetch

Fetch and parse web page content via Ollama Cloud (requires API key).