ConfigsMay 8, 2026·5 min read

Cherry Studio Custom Models — BYOK Any LLM Provider

Cherry Studio Custom Models adds any OpenAI-compatible endpoint — proxy, local, or third-party. Mix Claude, GPT, Gemini, DeepSeek, Ollama side-by-side.

Agent ready

This asset can be read and installed directly by agents

TokRepo exposes a universal CLI command, install contract, metadata JSON, adapter-aware plan, and raw content links so agents can judge fit, risk, and next actions.

Needs Confirmation · 66/100Policy: confirm
Agent surface
Any MCP/CLI agent
Kind
Skill
Install
Single
Trust
Trust: New
Entrypoint
Asset
Universal CLI install command
npx tokrepo install 121c2d1e-678a-4ea4-8799-70391f0a11de
Intro

Cherry Studio Custom Models support lets you add any OpenAI-compatible endpoint as a model — your own LiteLLM proxy, an Ollama instance, OpenRouter, DeepSeek, anything. Multiple models at once, swap mid-conversation, or run side-by-side comparisons. Best for: power users who want one interface across all their LLM providers, including local models. Works with: Cherry Studio 1.x, any OpenAI-compatible endpoint. Setup time: 2 minutes per model.


Add a custom provider

Settings → Models → ➕ Add Provider:

Provider Type: OpenAI Compatible
Name: My LiteLLM Proxy
Base URL: https://litellm.acme.internal/v1
API Key: sk-team-acme-xyz

Cherry Studio fetches the model list via /v1/models. Select the ones you want exposed.

Common providers

Provider Base URL Models
OpenRouter https://openrouter.ai/api/v1 300+
Together AI https://api.together.xyz/v1 200+
Groq https://api.groq.com/openai/v1 Fast Llama, Mixtral
DeepSeek https://api.deepseek.com/v1 DeepSeek-V3, R1, Coder
Ollama (local) http://localhost:11434/v1 Anything you've pulled
LM Studio (local) http://localhost:1234/v1 GGUF models
Your LiteLLM Proxy https://your-proxy.com/v1 Whatever it routes

Side-by-side comparison

In a chat, click the model selector → ⊕ Add Model. Two (or more) models respond to the same prompt simultaneously. Useful for picking the right model for a workload.

Switch mid-conversation

Click the model name on any message → switch the model used for the next message. Great for "use Sonnet for the hard reasoning, switch to Haiku for the boilerplate replies."

Config persistence

Provider configs are stored in ~/Library/Application Support/CherryStudio/config.json (encrypted with the OS keychain). Sync via the optional Cherry Studio Sync (encrypted to your passphrase) if you want them across devices.


FAQ

Q: Does it work with Anthropic Claude directly? A: Cherry Studio supports Anthropic's native API as a built-in provider type (not OpenAI-compatible). Add via Settings → Models → ➕ Add Provider → Anthropic. Same flow as the others.

Q: Can I use a self-hosted LLM? A: Yes — point to your local Ollama (port 11434) or LM Studio (port 1234) for fully local inference. The OpenAI-compatible adapter handles tool use and streaming for both.

Q: How are API keys stored? A: Locally only. Cherry Studio uses the OS keychain (macOS Keychain / Windows Credential Manager / Linux Secret Service). Keys never leave your device unless you opt into Cherry Studio Sync (encrypted).


Quick Use

  1. Open Cherry Studio → Settings → Models → ➕ Add Provider
  2. Pick OpenAI Compatible, paste base URL and API key
  3. Cherry Studio auto-fetches models via /v1/models — select the ones you want

Intro

Cherry Studio Custom Models support lets you add any OpenAI-compatible endpoint as a model — your own LiteLLM proxy, an Ollama instance, OpenRouter, DeepSeek, anything. Multiple models at once, swap mid-conversation, or run side-by-side comparisons. Best for: power users who want one interface across all their LLM providers, including local models. Works with: Cherry Studio 1.x, any OpenAI-compatible endpoint. Setup time: 2 minutes per model.


Add a custom provider

Settings → Models → ➕ Add Provider:

Provider Type: OpenAI Compatible
Name: My LiteLLM Proxy
Base URL: https://litellm.acme.internal/v1
API Key: sk-team-acme-xyz

Cherry Studio fetches the model list via /v1/models. Select the ones you want exposed.

Common providers

Provider Base URL Models
OpenRouter https://openrouter.ai/api/v1 300+
Together AI https://api.together.xyz/v1 200+
Groq https://api.groq.com/openai/v1 Fast Llama, Mixtral
DeepSeek https://api.deepseek.com/v1 DeepSeek-V3, R1, Coder
Ollama (local) http://localhost:11434/v1 Anything you've pulled
LM Studio (local) http://localhost:1234/v1 GGUF models
Your LiteLLM Proxy https://your-proxy.com/v1 Whatever it routes

Side-by-side comparison

In a chat, click the model selector → ⊕ Add Model. Two (or more) models respond to the same prompt simultaneously. Useful for picking the right model for a workload.

Switch mid-conversation

Click the model name on any message → switch the model used for the next message. Great for "use Sonnet for the hard reasoning, switch to Haiku for the boilerplate replies."

Config persistence

Provider configs are stored in ~/Library/Application Support/CherryStudio/config.json (encrypted with the OS keychain). Sync via the optional Cherry Studio Sync (encrypted to your passphrase) if you want them across devices.


FAQ

Q: Does it work with Anthropic Claude directly? A: Cherry Studio supports Anthropic's native API as a built-in provider type (not OpenAI-compatible). Add via Settings → Models → ➕ Add Provider → Anthropic. Same flow as the others.

Q: Can I use a self-hosted LLM? A: Yes — point to your local Ollama (port 11434) or LM Studio (port 1234) for fully local inference. The OpenAI-compatible adapter handles tool use and streaming for both.

Q: How are API keys stored? A: Locally only. Cherry Studio uses the OS keychain (macOS Keychain / Windows Credential Manager / Linux Secret Service). Keys never leave your device unless you opt into Cherry Studio Sync (encrypted).


Source & Thanks

Built by kangfenmao. Licensed under Apache-2.0.

CherryHQ/cherry-studio — ⭐ 18,000+

🙏

Source & Thanks

Built by kangfenmao. Licensed under Apache-2.0.

CherryHQ/cherry-studio — ⭐ 18,000+

Discussion

Sign in to join the discussion.
No comments yet. Be the first to share your thoughts.

Related Assets