ScriptsMar 30, 2026·2 min read

Open WebUI — Self-Hosted AI Chat Interface

User-friendly, self-hosted AI chat interface. Supports Ollama, OpenAI, Anthropic, and any OpenAI-compatible API. RAG, web search, voice, image gen, and plugins. 129K+ stars.

TL;DR
Self-hosted AI chat UI supporting Ollama, OpenAI, Anthropic, and any OpenAI-compatible API with RAG and plugins.
§01

What it is

Open WebUI is a self-hosted web interface for interacting with large language models. It provides a polished chat experience similar to ChatGPT but runs on your own infrastructure. The interface connects to Ollama for local models, OpenAI, Anthropic, and any API following the OpenAI-compatible format. Built-in features include RAG (retrieval-augmented generation), web search, voice input/output, image generation, and a plugin system.

This tool targets developers, teams, and organizations that want a private AI chat interface without sending data to third-party services. Self-hosters who already run Ollama benefit from a production-ready frontend.

§02

How it saves time or tokens

Open WebUI consolidates multiple LLM providers into a single interface, eliminating the need to switch between different chat UIs. The built-in RAG system lets you upload documents and query them directly, avoiding manual copy-paste of context into prompts. This reduces both token waste and the time spent formatting queries across different tools.

§03

How to use

  1. Deploy Open WebUI via Docker with a single command
  2. Connect it to your LLM backend (Ollama, OpenAI API key, or custom endpoint)
  3. Access the web UI in your browser and start chatting
§04

Example

# Quick start with Docker (connects to Ollama on the host)
docker run -d -p 3000:8080 \
  --add-host=host.docker.internal:host-gateway \
  -v open-webui:/app/backend/data \
  --name open-webui \
  --restart always \
  ghcr.io/open-webui/open-webui:main

# Access at http://localhost:3000
# First user to register becomes admin
§05

Related on TokRepo

  • Local LLM tools — Explore local model runners that pair with Open WebUI
  • RAG tools — Browse retrieval-augmented generation frameworks
§06

Common pitfalls

  • The default Docker setup expects Ollama on the host; use --add-host to bridge the network correctly
  • Uploading large documents for RAG requires sufficient disk space and may slow indexing on low-memory systems
  • Multi-user setups need proper authentication configuration to prevent unauthorized access to API keys

Frequently Asked Questions

Does Open WebUI work without Ollama?+

Yes. You can connect Open WebUI to any OpenAI-compatible API endpoint, including OpenAI directly, Anthropic via a proxy, or services like LiteLLM that unify multiple providers behind one API.

Is Open WebUI free to use?+

Yes. Open WebUI is open-source and free to self-host. There are no usage fees, subscription costs, or per-token charges from the UI itself. You only pay for the LLM provider you connect to.

How does the RAG feature work?+

Upload documents (PDF, text, markdown) through the web interface. Open WebUI indexes them locally and lets you query their contents in chat. The system retrieves relevant passages and includes them as context for the LLM.

Can multiple users share one Open WebUI instance?+

Yes. Open WebUI supports multi-user accounts with role-based access. The first registered user becomes admin. You can configure shared models, per-user API keys, and conversation privacy settings.

What hardware do I need to run Open WebUI?+

Open WebUI itself is lightweight and runs on any machine that supports Docker. The hardware requirements depend on your LLM backend. If using Ollama locally, you need sufficient RAM and optionally a GPU for the model you run.

Citations (3)
🙏

Source & Thanks

Created by Open WebUI. Licensed under BSD-3. open-webui/open-webui — 129,000+ GitHub stars

Discussion

Sign in to join the discussion.
No comments yet. Be the first to share your thoughts.

Related Assets