mcp-server-dify logo

mcp-server-dify

by yuru-sha

Model Context Protocol Server for Dify AI. This server enables LLMs to interact with Dify AI's chat completion capabilities through a standardized protocol.

View on GitHub

Last updated: N/A

What is mcp-server-dify?

mcp-server-dify is a Model Context Protocol server designed to integrate Large Language Models (LLMs) with Dify AI's chat completion API. It provides a standardized protocol for LLMs to interact with Dify AI, enabling context-aware conversations and access to Dify AI's features.

How to use mcp-server-dify?

To use the server, install it using NPM. Then, configure your application (e.g., Claude Desktop) to use the server by providing the Dify API endpoint and API key. The server can then be used to access tools like the meshi-doko restaurant recommendation tool by providing necessary parameters.

Key features of mcp-server-dify

  • Integration with Dify AI chat completion API

  • Restaurant recommendation tool (meshi-doko)

  • Support for conversation context

  • Streaming response support

Use cases of mcp-server-dify

  • Integrating LLMs with Dify AI for chat applications

  • Building context-aware conversational AI agents

  • Accessing Dify AI's tools and functionalities through a standardized protocol

  • Creating restaurant recommendation systems

FAQ from mcp-server-dify

What is the purpose of this server?

This server allows LLMs to interact with Dify AI's chat completion capabilities.

How do I install this server?

You can install it using NPM: npm install @modelcontextprotocol/server-dify

How do I configure this server?

You need to provide your Dify API endpoint and API key in the configuration.

What is meshi-doko?

meshi-doko is a restaurant recommendation tool that interfaces with Dify AI.

How can I contribute to this project?

You can submit a Pull Request with your contributions.