mcp-openai-gemini-llama-example

philschmid
171
A demo for building AI agents using MCP with open LLMs and SQLite.

Overview

What is mcp-openai-gemini-llama-example

mcp-openai-gemini-llama-example is a basic example project that demonstrates how to build an AI agent using the Model Context Protocol (MCP) with open LLMs like Meta Llama 3, OpenAI, or Google Gemini, along with a SQLite database. It serves as an educational tool rather than a production-ready framework.

How to Use

To use mcp-openai-gemini-llama-example, you need to have Docker installed and running, a Hugging Face account with an access token for the Llama 3 model, and a Google API key for the Gemini model. After cloning the repository and installing the required packages, you can run the agent in interactive mode to interact with the SQLite database.

Key Features

Key features include connecting to an MCP server, loading and using tools and resources from the MCP server, converting tools into LLM-compatible function calls, and interacting with LLMs using the openai SDK or google-genai SDK.

Where to Use

mcp-openai-gemini-llama-example can be used in educational settings for learning about AI agent development, database interactions, and the integration of different LLMs. It is suitable for developers and researchers interested in exploring AI technologies.

Use Cases

Use cases include building interactive AI agents that can query databases, providing responses based on user prompts, and demonstrating the capabilities of LLMs in real-time applications.

Content