Turbo MCP Client logo

Turbo MCP Client

by techspawn

Turbo MCP Client is a FastAPI application providing a web interface for interacting with Models Context Protocol (MCP) servers. It utilizes OpenAI's API for processing messages.

View on GitHub

Last updated: N/A

What is Turbo MCP Client?

This application is a web-based client that allows users to connect to multiple MCP servers, process messages through OpenAI's API, and interact with the system through a chat interface.

How to use Turbo MCP Client?

To use the application, clone the repository, install dependencies, configure environment variables and the config.json file, initialize the SQLite database, and then start the application using uvicorn main:app --reload. Access the application in your browser at http://localhost:8000 and enter your OpenAI API key.

Key features of Turbo MCP Client

  • Connect multiple servers simultaneously

  • Process messages through OpenAI's API

  • Interact via a web-based chat interface

  • Configure and store API license keys

Use cases of Turbo MCP Client

  • Interacting with language model servers

  • Building chat applications that leverage OpenAI

  • Managing configurations for multiple MCP server connections

  • Centralized message processing and management

FAQ from Turbo MCP Client

What is MCP?

MCP stands for Models Context Protocol. This application is based on this protocol to connect and work with model servers.

What is the required python version needed to use this repo?

The read me does not specify the version, but you can get the clues by looking at the python setup which will be on 3.x version and up.

What kind of models does this app use?

This app mainly uses Openai models like 'gpt-4o'. However, you can use any other open ai model that you have access too.

Where can I get the open ai license key from?

Please register to open ai platform.

Are contributions welcome?

Yes, contributions are always welcome. See contributing.md for ways to get started.