Browse and discover Model Context Protocol compatible clients.
The mcp-langchain-ts-client repository provides a LangChain.js client for the Model Context Protocol, porting the original Python implementation to JavaScript/TypeScript, enabling the use of MCP tools within LangChain.js applications.
Laravel MCP is a package that implements the Model Context Protocol, enabling communication between AI models and Laravel applications through standardized interfaces with features like tool registration and resource management.
This repository provides an MCP server that integrates Ollama with Claude Desktop, enabling users to list models, get model information, and ask questions to specified models using the Model Context Protocol.
This repository provides a TypeScript implementation of a Model Context Protocol (MCP) client, enabling LLM agents to interact with MCP servers through stdio or HTTP+SSE transports, supporting resources, tools, prompts, and sampling.
This repository provides two MCP client implementations that connect to an MCP server and send queries to Google Gemini AI. It enables Gemini to call external tools and maintains conversation
mcpx-py is a Python library that enables interaction with various Large Language Models (LLMs) using mcp.run tools, supporting models from PydanticAI, OpenAI, Ollama, and Gemini. mcpx-py is a Python library
MCPSwiftWrapper is a lightweight wrapper for mcp-swift-sdk, designed to integrate with LLM Swift libraries like SwiftOpenAI and SwiftAnthropic. It simplifies the creation of MCP Clients within macOS applications.
This repository showcases a simple Model Context Protocol (MCP) client using LangChain and TypeScript, demonstrating how to convert MCP server tools into LangChain-compatible tools for use with LLMs like Anthropic's Claude.
Desktop4mistral is a Python-based desktop application offering a user-friendly interface for interacting with Mistral AI models, featuring model selection, chat history, command support, and Markdown support. Desktop4mistral is a Python-based
This React-based demo showcases an MCP client interacting with SSE servers, enabling tool calls and text completion, though it's still under development and has limitations regarding tool naming and concurrent calls.
This repository provides a TypeScript-based Model Context Protocol (MCP) client that integrates with LangChain ReAct Agent, enabling interaction with LLMs like Anthropic, OpenAI, and Groq through MCP servers.
This repository provides a Python-based Model Context Protocol (MCP) client using LangChain, enabling interaction with MCP servers through LangChain ReAct Agent and supporting LLMs from Anthropic, OpenAI, and Groq.
This repository provides a LangChain client for the Model Context Protocol, enabling seamless connection to MCP servers. It converts MCP tools for use with LangChain, facilitating dynamic conversations.
Litemcp simplifies SDK adoption into MCP by providing a lightweight client for integrating AI SDKs like LangChain and Agent SDK, emphasizing simplicity, flexibility, and minimal dependencies. Litemcp simplifies SDK adoption
This repository provides an example client implementation using the Vercel AI SDK and Model Context Protocol SDK to streamline LLM chat interactions within a browser environment, including tool discovery and usage.