ConfigsApr 7, 2026·2 min read

Weaviate — AI-Native Vector Database

Open-source vector database for AI applications with built-in vectorization, hybrid search, and multi-tenancy. Supports 1B+ vectors with sub-100ms latency.

AI
AI Open Source · Community
Quick Use

Use it first, then decide how deep to go

This block should tell both the user and the agent what to copy, install, and apply first.

docker compose up -d
# or
pip install weaviate-client
import weaviate
import weaviate.classes as wvc

client = weaviate.connect_to_local()

# Create collection with auto-vectorization
collection = client.collections.create(
    name="Article",
    vectorizer_config=wvc.config.Configure.Vectorizer.text2vec_openai(),
)

# Add data (auto-vectorized)
collection.data.insert({"title": "AI trends", "body": "Vector databases are..."})

# Semantic search
results = collection.query.near_text(query="machine learning", limit=5)
for obj in results.objects:
    print(obj.properties["title"])

What is Weaviate?

Weaviate is an open-source, AI-native vector database designed for building AI applications at scale. It combines vector search with structured filtering, supports built-in vectorization from 20+ model providers, and handles billions of vectors with sub-100ms query latency.

Answer-Ready: Weaviate is an open-source AI-native vector database with built-in vectorization, hybrid search (vector + keyword), multi-tenancy, and support for 1B+ vectors at sub-100ms latency. Used by Stackexchange, Instabase, and Red Hat.

Best for: AI teams building RAG, semantic search, or recommendation systems. Works with: OpenAI, Cohere, Hugging Face, Claude (via embeddings). Setup time: Under 5 minutes with Docker.

Core Features

1. Built-In Vectorization

No need to manage embedding pipelines:

# Weaviate auto-vectorizes on insert
collection = client.collections.create(
    name="Document",
    vectorizer_config=wvc.config.Configure.Vectorizer.text2vec_cohere(),
    generative_config=wvc.config.Configure.Generative.anthropic(),
)

Supported providers: OpenAI, Cohere, Hugging Face, Google, AWS, Ollama, and more.

2. Hybrid Search

Combine vector similarity with keyword matching:

results = collection.query.hybrid(
    query="AI agent frameworks",
    alpha=0.75,  # 0=keyword only, 1=vector only
    limit=10,
)

3. Generative Search (RAG)

Query and generate in one step:

results = collection.generate.near_text(
    query="vector database comparison",
    single_prompt="Summarize this article in 2 sentences: {body}",
    limit=3,
)
for obj in results.objects:
    print(obj.generated)  # LLM-generated summary

4. Multi-Tenancy

Isolate data per tenant for SaaS applications:

collection = client.collections.create(
    name="UserDocs",
    multi_tenancy_config=wvc.config.Configure.multi_tenancy(enabled=True),
)
collection.tenants.create([wvc.tenants.Tenant(name="tenant_A")])

5. Filtering

results = collection.query.near_text(
    query="machine learning",
    filters=wvc.query.Filter.by_property("category").equal("research")
            & wvc.query.Filter.by_property("year").greater_than(2024),
    limit=5,
)

Deployment Options

Option Use Case
Docker Local development
Weaviate Cloud Managed production
Kubernetes Self-hosted at scale
Embedded In-process for testing

FAQ

Q: How does it compare to Pinecone? A: Weaviate is open-source and self-hostable with built-in vectorization. Pinecone is managed-only and requires external embedding.

Q: Can it handle production scale? A: Yes, supports 1B+ vectors with horizontal scaling and sub-100ms p99 latency.

Q: Does it support RAG out of the box? A: Yes, generative search combines retrieval and LLM generation in a single query.

🙏

Source & Thanks

Created by Weaviate. Licensed under BSD-3-Clause.

weaviate/weaviate — 12k+ stars

Discussion

Sign in to join the discussion.
No comments yet. Be the first to share your thoughts.