ConfigsApr 14, 2026·3 min read

Verba — The Golden RAGtriever by Weaviate

Verba is an open-source RAG (Retrieval-Augmented Generation) chatbot from the Weaviate team. Drop in PDFs, web pages, or notes; pick a model (OpenAI, Ollama, Anthropic); and get a polished chat UI with semantic search built in.

TL;DR
Verba bundles document ingestion, embedding, vector search, and chat into a single RAG application with a 2-minute setup.
§01

What it is

Verba is an open-source RAG (Retrieval-Augmented Generation) chatbot from the Weaviate team. It bundles document ingestion, chunking, embedding, vector search, and chat into a single application. Upload PDFs, DOCX, Markdown, plain text, web URLs, or GitHub repos, and Verba handles the entire pipeline from ingestion to conversational retrieval.

Verba is designed for teams who want to evaluate RAG patterns, demo to stakeholders, or self-host a private knowledge assistant without writing custom code.

§02

How it saves time or tokens

Building a RAG application from scratch requires choosing and integrating a vector store, embedding model, chunking strategy, and chat UI. Verba provides all of these in one package. The zero-infra mode runs Weaviate Embedded locally, so you do not need a separate database server. Pluggable components let you swap models and embedders without changing application code.

§03

How to use

  1. Install and run Verba:
pip install goldenverba
verba start
# Open http://localhost:8000
  1. Or run via Docker Compose:
git clone https://github.com/weaviate/Verba
cd Verba
docker compose up -d
  1. Upload documents through the web UI and start chatting with your knowledge base.
§04

Example

Configuring Verba to use Ollama for local, private RAG:

# Set environment variables before starting
export OLLAMA_URL='http://localhost:11434'
export OLLAMA_MODEL='llama3'
export OLLAMA_EMBED_MODEL='nomic-embed-text'

# Start Verba with Ollama as the backend
verba start

The architecture follows a pluggable pipeline:

[Document Upload] -> [Reader] -> [Chunker] -> [Embedder] -> [Weaviate Vector Store]
                                                                      |
[Chat UI] <- [LLM Generator] <- [Semantic Search] <------------------+
§05

Related on TokRepo

§06

Common pitfalls

  • Large PDF uploads without chunking configuration can exceed context windows; adjust chunk size in the settings for documents over 100 pages
  • The default Weaviate Embedded mode stores data in memory; restart loses all data unless you configure persistent storage
  • OpenAI API keys must be set as environment variables before starting; Verba does not prompt for them at runtime

Frequently Asked Questions

What LLM providers does Verba support?+

Verba supports OpenAI, Anthropic Claude, Ollama (for local models), OpenRouter, Cohere, and HuggingFace models. You configure the provider through environment variables or the web UI settings panel.

Can Verba run completely offline?+

Yes, with Ollama. Set Ollama as both the LLM and embedding provider, and Verba runs entirely on your local machine with no external API calls. This is useful for sensitive data that cannot leave your network.

What document formats does Verba accept?+

Verba supports PDF, DOCX, Markdown, plain text, web URLs, and GitHub repositories. The Reader component extracts text from each format before passing it to the chunking pipeline.

How is Verba different from LangChain?+

LangChain is a framework for building custom LLM applications. Verba is a complete, ready-to-use RAG application. If you need a quick RAG chatbot without coding, use Verba. If you need custom logic, pipelines, or integrations beyond chat, use LangChain.

Does Verba support multi-user access?+

Verba provides a single shared knowledge base by default. There is no built-in user authentication or per-user document separation. For multi-tenant deployments, you would need to run separate Verba instances or implement a reverse proxy with authentication.

Citations (3)
  • Verba GitHub— Verba is an open-source RAG application by Weaviate
  • Weaviate Docs— Weaviate vector database for semantic search
  • Anthropic Docs— RAG combines retrieval and generation for knowledge-grounded answers

Discussion

Sign in to join the discussion.
No comments yet. Be the first to share your thoughts.

Related Assets