ScriptsMar 31, 2026·2 min read

AnythingLLM — All-in-One AI Knowledge Base

All-in-one AI app: chat with documents, RAG, agents, multi-user, and 30+ LLM/embedding providers. Desktop + Docker. Privacy-first, no setup needed. 57K+ stars.

TL;DR
AnythingLLM bundles document chat, RAG, agents, and multi-user support into one privacy-first desktop or Docker app with 30+ LLM providers.
§01

What it is

AnythingLLM is an open-source, all-in-one AI application that lets you chat with your documents, build RAG pipelines, run AI agents, and manage multiple users -- all from a single desktop app or Docker container. It supports over 30 LLM and embedding providers including OpenAI, Anthropic, Ollama, and local models.

It targets non-technical users and small teams who want a ready-to-use AI knowledge base without writing code or managing infrastructure. Download the desktop app, drop in your documents, and start asking questions.

§02

How it saves time or tokens

AnythingLLM eliminates the need to build a custom RAG stack. Document ingestion, chunking, embedding, vector storage, and retrieval are all handled internally. You upload a PDF or a folder of files, and the system makes them searchable by the LLM. This is particularly valuable for teams who want document Q&A but lack the engineering resources to build a custom pipeline.

§03

How to use

  1. Download the desktop app from the official site, or run with Docker:
docker pull mintplexlabs/anythingllm
docker run -d -p 3001:3001 \
  -v anythingllm_storage:/app/server/storage \
  mintplexlabs/anythingllm
  1. Open the app and complete the setup wizard -- choose your LLM provider (OpenAI, Ollama, etc.) and embedding model.
  2. Create a workspace, upload documents (PDF, DOCX, TXT, CSV, and more).
  3. Start chatting. The system uses RAG to ground responses in your uploaded documents.
§04

Example

User: What were the key findings in the Q3 financial report?

AnythingLLM: Based on the uploaded Q3 report (pages 4-7),
revenue increased 12% year-over-year to $4.2M, driven
primarily by enterprise contract expansion...
[Source: Q3-Financial-Report.pdf, page 5]

Responses include source citations pointing back to the specific documents and pages.

§05

Related on TokRepo

§06

Common pitfalls

  • Document chunking settings affect answer quality significantly. The defaults work for general text, but code documentation or legal contracts may need custom chunk sizes and overlap settings.
  • The desktop app stores all data locally. If you need multi-user access, use the Docker deployment with authentication enabled.
  • LLM quality varies by provider. Using a small local model via Ollama works for simple Q&A, but complex multi-document reasoning benefits from a more capable model like GPT-4o or Claude.

Frequently Asked Questions

What document formats does AnythingLLM support?+

AnythingLLM supports PDF, DOCX, TXT, CSV, Markdown, and several other text formats. It can also ingest web pages by URL. Documents are automatically chunked and embedded for retrieval. Large documents are split into manageable pieces, and you can adjust chunk size and overlap in the workspace settings.

Can I use AnythingLLM with local models only?+

Yes. Connect AnythingLLM to a local Ollama server or any local LLM provider. Both the LLM and embedding model can run locally, meaning no data leaves your machine. This is the main appeal for users with privacy or compliance requirements. The desktop app makes this particularly easy to set up.

How does multi-user access work?+

In the Docker deployment, AnythingLLM supports multiple users with role-based access. Admins can create user accounts, assign them to workspaces, and control which documents each user can access. Each user has their own chat history. The desktop app is single-user by default.

Is AnythingLLM suitable for enterprise use?+

For small teams and departments, yes. It provides authentication, workspace isolation, and audit logging. For large enterprise deployments with hundreds of users, you may want to evaluate the managed cloud offering or consider a more scalable RAG platform. The open-source version has no user count limits but is designed for moderate scale.

What embedding models does AnythingLLM support?+

AnythingLLM supports embedding models from OpenAI, Cohere, local models via Ollama, and several other providers. You configure the embedding model in the workspace settings. The choice of embedding model affects retrieval quality -- OpenAI and Cohere embeddings generally provide strong results, while local alternatives trade some quality for privacy.

Citations (3)
🙏

Source & Thanks

Created by Mintplex Labs. Licensed under MIT. Mintplex-Labs/anything-llm — 57,000+ GitHub stars

Discussion

Sign in to join the discussion.
No comments yet. Be the first to share your thoughts.

Related Assets