Skip to content

Commit 8b58386

Browse files
authored
docs: clarify LLM API configuration in mcp_simple_chatbot README (#487)
Co-authored-by: Danny <88821366+ddaltn@users.noreply.github.com>
1 parent 173e0ee commit 8b58386

File tree

1 file changed

+1
-0
lines changed

1 file changed

+1
-0
lines changed

examples/clients/simple-chatbot/README.MD

+1
Original file line numberDiff line numberDiff line change
@@ -25,6 +25,7 @@ This example demonstrates how to integrate the Model Context Protocol (MCP) into
2525
```plaintext
2626
LLM_API_KEY=your_api_key_here
2727
```
28+
**Note:** The current implementation is configured to use the Groq API endpoint (`https://api.groq.com/openai/v1/chat/completions`) with the `llama-3.2-90b-vision-preview` model. If you plan to use a different LLM provider, you'll need to modify the `LLMClient` class in `main.py` to use the appropriate endpoint URL and model parameters.
2829

2930
3. **Configure servers:**
3031

0 commit comments

Comments
 (0)