An MCP server implemented with fastmcp, OpenAI embeddings, and qdrant-client, providing long-term memory and semantic search on top of Qdrant.
- MCP server built with
fastmcp - Hybrid search in Qdrant (dense OpenAI embeddings + sparse BM25)
- Chinese support via jieba
- Knowledge Base tools (renamed from memory tools to avoid conflict):
store-knowledgestore-knowledge-bulksearch-knowledgeget-knowledge-by-idinspect-knowledge-basedelete-knowledge
- Multiple transports: stdio, SSE, streamable HTTP
- Python 3.12+
- Qdrant reachable via HTTP
The project is published as the better-qdrant-mcp package, so you can run it directly with uvx without cloning this repo.
Minimal env for typical use:
QDRANT_URL– defaults tohttp://localhost:6333QDRANT_API_KEY– optionalCOLLECTION_NAME– optional default collectionOPENAI_API_KEY(orOPENAPI_API_KEY) – requiredOPENAI_BASE_URL– optionalOPENAI_EMBEDDING_MODEL– defaults totext-embedding-3-small
Advanced / transport-related env:
MCP_TRANSPORT–stdio|sse|streamable-http(default:stdio)MCP_HOST– host for HTTP-based transports (default:0.0.0.0)MCP_PORT– port for HTTP-based transports (default:8000)MCP_PATH– path for HTTP transports (default:/mcp)
Once the server is running, the MCP client will see these tools:
store-knowledge(content: str, title?: str, tags?: list[str], metadata?: dict, collection_name?: str) -> strstore-knowledge-bulk(items: list[KnowledgeItem], collection_name?: str) -> strsearch-knowledge(query: str, limit?: int=5, collection_name?: str) -> strget-knowledge-by-id(ids: list[str] | str, collection_name?: str) -> strinspect-knowledge-base(collection_name?: str) -> strdelete-knowledge(ids: list[str] | str, collection_name?: str) -> str
store-knowledge automatically embeds the text using OpenAI and stores it in Qdrant (Knowledge Base), returning the stored point ID. The title and tags fields help improve search context and categorization.
store-knowledge-bulk efficiently stores multiple knowledge items at once using batch embedding. Each item in the list should include content (required), and optionally title, tags, and metadata fields. This is more efficient than calling store-knowledge multiple times.
search-knowledge uses hybrid search in Qdrant (dense + sparse). If the collection is configured with named vectors dense and sparse, queries are ranked by fusing dense OpenAI embeddings and sparse BM25 scores; otherwise it falls back to dense-only search.
get-knowledge-by-id retrieves the complete payload information for one or more knowledge items by their point IDs. Use this to inspect the full details of stored items (including content, title, tags, metadata, and stored_at timestamp). You can pass a single ID or a list of IDs (typically using the id field returned by search-knowledge).
inspect-knowledge-base shows the collection configuration and sample data points, useful for debugging and verification.
delete-knowledge deletes one or more stored knowledge items from Qdrant by their point IDs. You can pass a single ID or a list of IDs (typically using the id field returned by search-knowledge).
You can either specify the transport via CLI flags (recommended for quick start) or via env (MCP_TRANSPORT).
uvx better-qdrant-mcpIn this mode, you configure your MCP client to use stdio transport and just invoke the binary; no HTTP URL is needed.
# Default host 0.0.0.0 and port 8000
uvx better-qdrant-mcp --transport sse
# Custom host and port
uvx better-qdrant-mcp --transport sse --host 0.0.0.0 --port 3000Connection details for MCP clients:
- Transport:
sse - URL:
http://<host>:<port>/sse(for example:http://localhost:8000/sse)
# Default host 0.0.0.0, port 8000 and path /mcp
uvx better-qdrant-mcp --transport streamable-http
# Custom host, port, and path
uvx better-qdrant-mcp --transport streamable-http --host 0.0.0.0 --port 3000 --path /api/mcpConnection details for MCP clients:
- Transport:
streamable-http - URL:
http://<host>:<port><path>(for example:http://localhost:8000/mcp)
If you want to work on this repo locally instead of using the published package:
# using uv (recommended)
uv sync
# or with pip (editable install)
pip install -e .For local development, you can use the provided Makefile:
make buildThis command will first clean the dist directory and then run uv build to produce fresh artifacts.
Docker Compose provides Qdrant + this MCP server as a single service. Transport (stdio, sse, streamable-http) is selected via MCP_TRANSPORT.
The Docker image is built and published automatically to GitHub Container Registry as:
ghcr.io/jtsang4/better-qdrant-mcp:latest- Additional tags for branches, tags, and commit SHAs
The provided docker-compose.yml uses this published image directly, so you do not need to build the image locally.
# Start Qdrant + MCP using the published image
docker compose up -d
# Pull the latest published image and restart services
docker compose pull mcp && docker compose up -d
# Stop services
docker compose down