This repository provides an MCP server for Kibela API integration, enabling Large Language Models to search notes, retrieve recent notes, and access note content with comments, facilitating interaction with Kibela content.
This repository provides an MCP (Model Context Protocol) server implementation designed to integrate LLMs (Large Language Models) with the Kibela API. This integration allows LLMs to interact with and retrieve content from Kibela, a knowledge-sharing platform.
<img width="320" alt="Example" src="https://github.com/user-attachments/assets/eeed8f45-eb24-456d-bb70-9e738aa1bfb3" /><a href="https://glama.ai/mcp/servers/m21nkeig1p"><img width="380" height="200" src="https://glama.ai/mcp/servers/m21nkeig1p/badge" alt="Kibela Server MCP server" /></a>
The server requires KIBELA_TEAM
and KIBELA_TOKEN
environment variables. Configuration examples are provided for Claude Desktop and Cursor, detailing how to integrate the server by specifying the command, arguments, and environment variables within their respective configuration files. For Cursor integration using SSE, the server URL should be set to http://localhost:3000/sse
.
The server exposes the following tools: kibela_search_notes
(search Kibela notes), kibela_get_my_notes
(get latest notes), and kibela_get_note_content
(get note content and comments). Each tool's input parameters and return values are documented.
kiwamizamurai/mcp-kibela-server
February 1, 2025
February 11, 2025
TypeScript