MCP ConfigsApr 6, 2026·2 min read

Qdrant MCP — Vector Search Engine for AI Agents

MCP server for Qdrant vector database. Gives AI agents the power to store and search embeddings for RAG, semantic search, and recommendation systems. 22,000+ stars on Qdrant.

TL;DR
MCP server for Qdrant vector database enabling AI agents to store and search embeddings for RAG and semantic search.
§01

What it is

Qdrant MCP is a Model Context Protocol server for the Qdrant vector database. It gives AI agents the ability to store, search, and manage vector embeddings directly through MCP tool calls. This enables agents to build and query knowledge bases, implement semantic search, and power recommendation systems without the developer writing database integration code.

The project targets developers building AI agents that need persistent vector storage -- RAG systems, semantic search engines, and recommendation pipelines. The MCP transport means any compatible AI client can use Qdrant without custom API wrappers.

§02

How it saves time or tokens

Integrating a vector database into an AI agent workflow typically requires writing embedding generation, database client setup, collection management, and query construction code. Qdrant MCP packages all of this into MCP tools that the agent calls natively. The agent can store embeddings, search by similarity, and manage collections through natural language commands.

Semantic search via Qdrant also reduces tokens by retrieving only the most relevant documents rather than loading entire knowledge bases into the context window.

§03

How to use

  1. Start a Qdrant instance:
docker run -p 6333:6333 qdrant/qdrant
  1. Add to your MCP config:
{
  "mcpServers": {
    "qdrant": {
      "command": "npx",
      "args": ["-y", "@qdrant/mcp-server"],
      "env": {
        "QDRANT_URL": "http://localhost:6333"
      }
    }
  }
}
  1. Restart your AI tool and start storing and searching vectors.
§04

Example

User: Store the contents of our API documentation in Qdrant
      for semantic search.

Agent (via Qdrant MCP):
  1. Created collection 'api_docs' with 1536-dimension vectors
  2. Embedded 47 documentation pages
  3. Stored all embeddings with metadata

User: Find documentation related to rate limiting.

Agent (via Qdrant MCP):
  Search results (top 3):
  1. Rate Limiting Guide (score: 0.94)
  2. API Authentication (score: 0.72) - mentions rate limits per key
  3. Error Handling (score: 0.68) - covers 429 responses

The agent manages the full vector lifecycle through MCP tools.

§05

Related on TokRepo

§06

Common pitfalls

  • Forgetting to start the Qdrant server before connecting -- the MCP server needs a running Qdrant instance to connect to.
  • Using mismatched vector dimensions -- the embedding model's output dimension must match the collection's configured dimension. Check your embedding model's documentation.
  • Not setting up persistent storage for Qdrant -- the default Docker setup may not persist data. Mount a volume for production use.

Frequently Asked Questions

Do I need a Qdrant Cloud account?+

No. You can run Qdrant locally via Docker. The MCP server connects to any Qdrant instance, whether local or cloud-hosted. For production use, Qdrant Cloud offers managed hosting with scaling and backups.

Which embedding models work with Qdrant MCP?+

Qdrant MCP is embedding-model agnostic. You can use OpenAI embeddings, Cohere, or any other embedding provider. The MCP server stores whatever vectors you provide -- the embedding generation can happen in the agent or a separate service.

Can AI agents create and manage collections?+

Yes. The MCP tools include collection management operations -- creating collections with specified dimensions, listing existing collections, and deleting collections. The agent handles the full lifecycle through natural language.

Is Qdrant suitable for production RAG systems?+

Yes. Qdrant is a mature vector database used in production by many companies. It supports horizontal scaling, replication, filtering, and payload indexing. The MCP server adds an AI agent interface on top of Qdrant's production-grade engine.

How does Qdrant MCP handle large document sets?+

For large document sets, you chunk documents before embedding and store each chunk as a separate point in Qdrant. The MCP server handles batch insertions efficiently. Qdrant itself is designed for billions of vectors.

Citations (3)
🙏

Source & Thanks

Created by Qdrant. Licensed under Apache 2.0.

qdrant — ⭐ 22,000+

Thanks to the Qdrant team for bringing vector search to AI agents.

Discussion

Sign in to join the discussion.
No comments yet. Be the first to share your thoughts.