Skip to content

Commit 059e831

Browse files
authored
Merge pull request #98 from 3choff/feature/simple-chatbot-example
Add example client to examples/clients folder
2 parents 0cb3b26 + eba1ad2 commit 059e831

File tree

8 files changed

+601
-0
lines changed

8 files changed

+601
-0
lines changed
Original file line numberDiff line numberDiff line change
@@ -0,0 +1 @@
1+
3.10
+110
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,110 @@
1+
# MCP Simple Chatbot
2+
3+
This example demonstrates how to integrate the Model Context Protocol (MCP) into a simple CLI chatbot. The implementation showcases MCP's flexibility by supporting multiple tools through MCP servers and is compatible with any LLM provider that follows OpenAI API standards.
4+
5+
## Requirements
6+
7+
- Python 3.10
8+
- `python-dotenv`
9+
- `requests`
10+
- `mcp`
11+
- `uvicorn`
12+
13+
## Installation
14+
15+
1. **Install the dependencies:**
16+
17+
```bash
18+
pip install -r requirements.txt
19+
```
20+
21+
2. **Set up environment variables:**
22+
23+
Create a `.env` file in the root directory and add your API key:
24+
25+
```plaintext
26+
LLM_API_KEY=your_api_key_here
27+
```
28+
29+
3. **Configure servers:**
30+
31+
The `servers_config.json` follows the same structure as Claude Desktop, allowing for easy integration of multiple servers.
32+
Here's an example:
33+
34+
```json
35+
{
36+
"mcpServers": {
37+
"sqlite": {
38+
"command": "uvx",
39+
"args": ["mcp-server-sqlite", "--db-path", "./test.db"]
40+
},
41+
"puppeteer": {
42+
"command": "npx",
43+
"args": ["-y", "@modelcontextprotocol/server-puppeteer"]
44+
}
45+
}
46+
}
47+
```
48+
Environment variables are supported as well. Pass them as you would with the Claude Desktop App.
49+
50+
Example:
51+
```json
52+
{
53+
"mcpServers": {
54+
"server_name": {
55+
"command": "uvx",
56+
"args": ["mcp-server-name", "--additional-args"],
57+
"env": {
58+
"API_KEY": "your_api_key_here"
59+
}
60+
}
61+
}
62+
}
63+
```
64+
65+
## Usage
66+
67+
1. **Run the client:**
68+
69+
```bash
70+
python main.py
71+
```
72+
73+
2. **Interact with the assistant:**
74+
75+
The assistant will automatically detect available tools and can respond to queries based on the tools provided by the configured servers.
76+
77+
3. **Exit the session:**
78+
79+
Type `quit` or `exit` to end the session.
80+
81+
## Architecture
82+
83+
- **Tool Discovery**: Tools are automatically discovered from configured servers.
84+
- **System Prompt**: Tools are dynamically included in the system prompt, allowing the LLM to understand available capabilities.
85+
- **Server Integration**: Supports any MCP-compatible server, tested with various server implementations including Uvicorn and Node.js.
86+
87+
### Class Structure
88+
- **Configuration**: Manages environment variables and server configurations
89+
- **Server**: Handles MCP server initialization, tool discovery, and execution
90+
- **Tool**: Represents individual tools with their properties and formatting
91+
- **LLMClient**: Manages communication with the LLM provider
92+
- **ChatSession**: Orchestrates the interaction between user, LLM, and tools
93+
94+
### Logic Flow
95+
96+
1. **Tool Integration**:
97+
- Tools are dynamically discovered from MCP servers
98+
- Tool descriptions are automatically included in system prompt
99+
- Tool execution is handled through standardized MCP protocol
100+
101+
2. **Runtime Flow**:
102+
- User input is received
103+
- Input is sent to LLM with context of available tools
104+
- LLM response is parsed:
105+
- If it's a tool call → execute tool and return result
106+
- If it's a direct response → return to user
107+
- Tool results are sent back to LLM for interpretation
108+
- Final response is presented to user
109+
110+
Original file line numberDiff line numberDiff line change
@@ -0,0 +1 @@
1+
GROQ_API_KEY=gsk_1234567890

0 commit comments

Comments
 (0)