ScriptsApr 1, 2026·1 min read

Quivr — Opinionated RAG Framework for Any LLM

Quivr is an opinionated RAG framework supporting any LLM, multiple file types, and customizable retrieval. 39.1K+ stars. Apache 2.0.

TL;DR
Opinionated RAG framework that builds a queryable knowledge brain from documents, supporting any LLM backend.
§01

What it is

Quivr is a Python RAG (retrieval-augmented generation) framework that takes an opinionated approach to building knowledge bases from documents. You feed it files (PDFs, text, markdown, and more), and Quivr handles ingestion, chunking, embedding, and retrieval. Then you query your document collection in natural language using any LLM backend. The framework is designed to get a working RAG pipeline running in minutes, not days.

Developers and teams who need to build document Q&A systems, internal knowledge bases, or AI assistants grounded in specific documents benefit from Quivr. Its opinionated defaults mean less configuration compared to more modular frameworks.

§02

How it saves time or tokens

Quivr's opinionated design eliminates the decision fatigue of choosing chunking strategies, embedding models, and retrieval methods. The defaults work well for most document types. By handling the entire RAG pipeline in a few lines of code, Quivr saves the days of setup that more flexible frameworks require. The token_estimate for this workflow is approximately 337 tokens for a basic query.

§03

How to use

  1. Install Quivr via pip
  2. Create a Brain from your document files
  3. Ask questions in natural language and get grounded answers
§04

Example

from quivr_core import Brain

# Create a brain from your documents
brain = Brain.from_files(
    name='my-knowledge-base',
    file_paths=['./report.pdf', './notes.md']
)

# Ask questions
answer = brain.ask('What were the key findings?')
print(answer.answer)
print(answer.sources)  # Shows which documents were used
§05

Related on TokRepo

§06

Common pitfalls

  • Large PDF files with complex layouts may chunk poorly; preprocess scanned documents with OCR before ingestion
  • The default embedding model requires an API key; configure a local embedding model for fully offline operation
  • Quivr's opinionated defaults work well for general documents but may need tuning for highly specialized technical content

Frequently Asked Questions

Which LLMs does Quivr support?+

Quivr supports any LLM that provides a chat API, including OpenAI, Anthropic Claude, Mistral, and local models via Ollama. You configure the LLM backend when creating or querying a Brain.

What file types can Quivr ingest?+

Quivr handles PDFs, markdown, plain text, Word documents, and several other formats. The framework includes parsers for each type and chunks them appropriately for retrieval.

How is Quivr different from LangChain RAG?+

LangChain is modular and requires you to wire together components manually. Quivr is opinionated and provides sensible defaults for the entire pipeline. Quivr gets you running faster; LangChain gives you more control.

Can I use Quivr without an API key?+

You need an API key for the LLM and embedding model by default. To run fully offline, configure Quivr to use a local LLM via Ollama and a local embedding model. This removes all external API dependencies.

Does Quivr support multi-user access?+

Quivr Core is a library for single-user programmatic use. The Quivr platform (separate project) adds multi-user support with a web UI, authentication, and shared brains.

Citations (3)
🙏

Source & Thanks

QuivrHQ/quivr — 39,100+ GitHub stars

Discussion

Sign in to join the discussion.
No comments yet. Be the first to share your thoughts.

Related Assets