Open WebUI — Self-Hosted AI Chat Interface
User-friendly, self-hosted AI chat interface. Supports Ollama, OpenAI, Anthropic, and any OpenAI-compatible API. RAG, web search, voice, image gen, and plugins. 129K+ stars.
What it is
Open WebUI is a self-hosted web interface for interacting with large language models. It provides a polished chat experience similar to ChatGPT but runs on your own infrastructure. The interface connects to Ollama for local models, OpenAI, Anthropic, and any API following the OpenAI-compatible format. Built-in features include RAG (retrieval-augmented generation), web search, voice input/output, image generation, and a plugin system.
This tool targets developers, teams, and organizations that want a private AI chat interface without sending data to third-party services. Self-hosters who already run Ollama benefit from a production-ready frontend.
How it saves time or tokens
Open WebUI consolidates multiple LLM providers into a single interface, eliminating the need to switch between different chat UIs. The built-in RAG system lets you upload documents and query them directly, avoiding manual copy-paste of context into prompts. This reduces both token waste and the time spent formatting queries across different tools.
How to use
- Deploy Open WebUI via Docker with a single command
- Connect it to your LLM backend (Ollama, OpenAI API key, or custom endpoint)
- Access the web UI in your browser and start chatting
Example
# Quick start with Docker (connects to Ollama on the host)
docker run -d -p 3000:8080 \
--add-host=host.docker.internal:host-gateway \
-v open-webui:/app/backend/data \
--name open-webui \
--restart always \
ghcr.io/open-webui/open-webui:main
# Access at http://localhost:3000
# First user to register becomes admin
Related on TokRepo
- Local LLM tools — Explore local model runners that pair with Open WebUI
- RAG tools — Browse retrieval-augmented generation frameworks
Common pitfalls
- The default Docker setup expects Ollama on the host; use
--add-hostto bridge the network correctly - Uploading large documents for RAG requires sufficient disk space and may slow indexing on low-memory systems
- Multi-user setups need proper authentication configuration to prevent unauthorized access to API keys
Frequently Asked Questions
Yes. You can connect Open WebUI to any OpenAI-compatible API endpoint, including OpenAI directly, Anthropic via a proxy, or services like LiteLLM that unify multiple providers behind one API.
Yes. Open WebUI is open-source and free to self-host. There are no usage fees, subscription costs, or per-token charges from the UI itself. You only pay for the LLM provider you connect to.
Upload documents (PDF, text, markdown) through the web interface. Open WebUI indexes them locally and lets you query their contents in chat. The system retrieves relevant passages and includes them as context for the LLM.
Yes. Open WebUI supports multi-user accounts with role-based access. The first registered user becomes admin. You can configure shared models, per-user API keys, and conversation privacy settings.
Open WebUI itself is lightweight and runs on any machine that supports Docker. The hardware requirements depend on your LLM backend. If using Ollama locally, you need sufficient RAM and optionally a GPU for the model you run.
Citations (3)
- Open WebUI GitHub— Self-hosted AI chat interface with RAG, web search, voice, and plugins
- Open WebUI Documentation— Supports Ollama, OpenAI, Anthropic, and OpenAI-compatible APIs
- Open WebUI Getting Started— Docker deployment with multi-user support
Related on TokRepo
Source & Thanks
Created by Open WebUI. Licensed under BSD-3. open-webui/open-webui — 129,000+ GitHub stars
Discussion
Related Assets
WCDB — WeChat Cross-Platform Database Framework
A high-performance, cross-platform database framework developed by WeChat, built on SQLite with ORM, encryption, repair, and migration capabilities.
sql.js — Run SQLite in the Browser with WebAssembly
A JavaScript library that compiles SQLite to WebAssembly, letting you run a full SQL database entirely in the browser or Node.js.
Realm — High-Performance Mobile Database
A fast, object-oriented mobile database designed as a modern replacement for SQLite and Core Data on iOS and Android.