Quick Use
- Open Cherry Studio → Settings → Models → ➕ Add Provider
- Pick OpenAI Compatible, paste base URL and API key
- Cherry Studio auto-fetches models via /v1/models — select the ones you want
Intro
Cherry Studio Custom Models support lets you add any OpenAI-compatible endpoint as a model — your own LiteLLM proxy, an Ollama instance, OpenRouter, DeepSeek, anything. Multiple models at once, swap mid-conversation, or run side-by-side comparisons. Best for: power users who want one interface across all their LLM providers, including local models. Works with: Cherry Studio 1.x, any OpenAI-compatible endpoint. Setup time: 2 minutes per model.
Add a custom provider
Settings → Models → ➕ Add Provider:
Provider Type: OpenAI Compatible
Name: My LiteLLM Proxy
Base URL: https://litellm.acme.internal/v1
API Key: sk-team-acme-xyzCherry Studio fetches the model list via /v1/models. Select the ones you want exposed.
Common providers
| Provider | Base URL | Models |
|---|---|---|
| OpenRouter | https://openrouter.ai/api/v1 |
300+ |
| Together AI | https://api.together.xyz/v1 |
200+ |
| Groq | https://api.groq.com/openai/v1 |
Fast Llama, Mixtral |
| DeepSeek | https://api.deepseek.com/v1 |
DeepSeek-V3, R1, Coder |
| Ollama (local) | http://localhost:11434/v1 |
Anything you've pulled |
| LM Studio (local) | http://localhost:1234/v1 |
GGUF models |
| Your LiteLLM Proxy | https://your-proxy.com/v1 |
Whatever it routes |
Side-by-side comparison
In a chat, click the model selector → ⊕ Add Model. Two (or more) models respond to the same prompt simultaneously. Useful for picking the right model for a workload.
Switch mid-conversation
Click the model name on any message → switch the model used for the next message. Great for "use Sonnet for the hard reasoning, switch to Haiku for the boilerplate replies."
Config persistence
Provider configs are stored in ~/Library/Application Support/CherryStudio/config.json (encrypted with the OS keychain). Sync via the optional Cherry Studio Sync (encrypted to your passphrase) if you want them across devices.
FAQ
Q: Does it work with Anthropic Claude directly? A: Cherry Studio supports Anthropic's native API as a built-in provider type (not OpenAI-compatible). Add via Settings → Models → ➕ Add Provider → Anthropic. Same flow as the others.
Q: Can I use a self-hosted LLM? A: Yes — point to your local Ollama (port 11434) or LM Studio (port 1234) for fully local inference. The OpenAI-compatible adapter handles tool use and streaming for both.
Q: How are API keys stored? A: Locally only. Cherry Studio uses the OS keychain (macOS Keychain / Windows Credential Manager / Linux Secret Service). Keys never leave your device unless you opt into Cherry Studio Sync (encrypted).
Source & Thanks
Built by kangfenmao. Licensed under Apache-2.0.
CherryHQ/cherry-studio — ⭐ 18,000+