ollama-MCP-server logo

ollama-MCP-server

by NewAITees

The ollama-MCP-server enables seamless integration between local Ollama LLM instances and MCP-compatible applications. It provides advanced task decomposition, evaluation, and workflow management.

View on GitHub

Last updated: N/A

What is ollama-MCP-server?

The ollama-MCP-server is a Model Context Protocol (MCP) server that communicates with Ollama. It allows MCP-compatible applications to interact with local Ollama LLM instances for advanced task decomposition, evaluation, and workflow management.

How to use ollama-MCP-server?

To use the server, install it via pip (pip install ollama-mcp-server). Configure your Claude Desktop (or other MCP client) with the appropriate settings, specifying the command and arguments to run the server. Use the provided tools (decompose-task, evaluate-result, run-model) via the MCP protocol, sending requests with the required parameters.

Key features of ollama-MCP-server

  • Task decomposition for complex problems

  • Result evaluation and validation

  • Ollama model management and execution

  • Standardized communication via MCP protocol

  • Enhanced error handling with detailed messages

  • Performance optimization (connection pooling, LRU cache)

Use cases of ollama-MCP-server

  • Breaking down complex tasks into manageable subtasks

  • Evaluating the results of LLM-generated content against specific criteria

  • Running Ollama models with specific prompts and parameters

  • Integrating LLMs into MCP-compatible applications for automated workflows

FAQ from ollama-MCP-server

What is the purpose of the task:// resource?

The task:// resource provides access to individual tasks, allowing you to manage and interact with them.

How does the server handle errors?

The server provides detailed and structured error messages, including a message, status code, and details about the error.

What are the benefits of connection pooling?

Connection pooling improves request performance and reduces resource usage by reusing HTTP connections.

How can I specify which Ollama model to use?

You can specify the model via the tool call parameters, the MCP configuration file, an environment variable (OLLAMA_DEFAULT_MODEL), or it will default to llama3.

How do I run the tests?

Use the ./run_tests.sh script with options like --unit, --integration, or --all to run different test suites.

ollama-MCP-server - MCP Server | MCP Directory