Skip to content

couchbase-examples/mcp-cookbook

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

34 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Building AI Agents using Couchbase MCP Server

This project showcases building AI Agents using Couchbase with the official Couchbase Model Context Protocol (MCP) server.

Users can ask questions like:

  • "List out the top 5 hotels by the highest aggregate rating?"
  • "Recommend me a flight and hotel from New York to San Francisco."
  • And other questions relevant to the specific dataset and agent implementation in each tutorial.

Introduction to Model Context Protocol (MCP)

The Model Context Protocol (MCP) is an open standard designed to standardize how AI assistants and applications connect to and interact with external data sources, tools, and systems. Think of MCP as a universal adapter that allows AI models to seamlessly access the context they need to produce more relevant, accurate, and actionable responses.

Key Goals and Features of MCP:

  • Standardized Communication: MCP provides a common language and structure for AI models to communicate with diverse backend systems, replacing the need for numerous custom integrations.
  • Enhanced Context Management: It helps manage the limited context windows of LLMs efficiently, enabling them to maintain longer, more coherent interactions and leverage historical data.
  • Secure Data Access: MCP emphasizes secure connections, allowing developers to expose data through MCP servers while maintaining control over their infrastructure.
  • Tool Use and Actionability: It enables LLMs to not just retrieve information but also to use external tools and trigger actions in other systems.
  • Interoperability: Fosters an ecosystem where different AI tools, models, and data sources can work together more cohesively.
  • Client-Server Architecture: MCP typically involves:
    • MCP Hosts/Clients: Applications (like AI assistants, IDEs, or other AI-powered tools) that want to access data or capabilities. In some demos, LangChain acts as an MCP client.
    • MCP Servers: Lightweight programs that expose specific data sources or tools (e.g., a database, an API) through the standardized MCP. The mcp-server-couchbase project fulfills this role for Couchbase.

MCP aims to break down data silos, making it easier for AI to integrate with real-world applications and enterprise systems, leading to more powerful and context-aware AI solutions. It is an open-source initiative, often supported by SDKs in various programming languages to facilitate the development of MCP clients and servers.

Prerequisites

  • Python 3.10 or higher.
  • A running Couchbase cluster. The easiest way to get started is to use Capella free tier, which is fully managed version of Couchbase server. You can follow instructions to import one of the sample datasets or import your own.
  • uv installed to run the MCP server via uvx.

Running the MCP Server

The tutorials in this cookbook use the prebuilt couchbase-mcp-server package from PyPI. There are two ways to run the MCP server:

Option 1: Using Prebuilt Package (Recommended)

The easiest way is to use the prebuilt package via uvx:

uvx couchbase-mcp-server

This automatically downloads and executes the latest version of the MCP server from PyPI. The server will be started with environment variables from your .env file or system environment.

Option 2: Running from Source

If you need to modify the server or run a specific version, you can clone and run from source:

git clone https://github.com/Couchbase-Ecosystem/mcp-server-couchbase.git
cd mcp-server-couchbase

you can execute server script (src/mcp_server.py) using uv (Python project manager).

Note: The tutorials in this cookbook are configured to use the prebuilt package via uvx. If you choose to run from source, you'll need to update the server configuration in the notebooks accordingly.

For more details about the MCP server, visit the mcp-server-couchbase GitHub repository.

Setup

1. Clone the repository:

git clone https://github.com/couchbase-examples/mcp-cookbook.git
cd mcp-cookbook

2. Fill in environment variables:

Copy the .env.sample file in each tutorial's directory to .env and fill in the environment variables:

# Example for LangChain tutorial
cp langchain_mcp_adapter/.env.sample langchain_mcp_adapter/.env

3. Install Dependencies

Each tutorial notebook includes a cell with pinned dependency versions. The notebooks will install these dependencies automatically when you run the installation cell.

4. Run the notebook file

You can either run the notebook file on Google Colab or run it on your system by setting up the Python environment.

Support Policy

Contributions are welcome! Please feel free to submit a pull request or open an issue for any bugs or feature requests.

License

This project is licensed under the MIT License. See the LICENSE file for details.

About

No description, website, or topics provided.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors