MCP ConfigsApr 4, 2026·3 min read

Pal MCP Server — Multi-Model AI Gateway for Claude Code

MCP server that lets Claude Code use Gemini, OpenAI, Grok, and Ollama as a unified AI dev team. Features model routing, CLI-to-CLI bridge, and conversation continuity across 7+ providers.

TO
TokRepo精选 · Community
Quick Use

Use it first, then decide how deep to go

This block should tell both the user and the agent what to copy, install, and apply first.

Add to your .mcp.json:

{
  "mcpServers": {
    "pal": {
      "command": "bash",
      "args": ["-c", "for p in $(which uvx 2>/dev/null) $HOME/.local/bin/uvx /opt/homebrew/bin/uvx /usr/local/bin/uvx uvx; do [ -x \"$p\" ] && exec \"$p\" --from git+https://github.com/BeehiveInnovations/pal-mcp-server.git pal-mcp-server; done; echo 'uvx not found' >&2; exit 1"],
      "env": {
        "GEMINI_API_KEY": "your-gemini-key",
        "DEFAULT_MODEL": "auto"
      }
    }
  }
}

Prerequisites: Python 3.10+, Git, uv (pip install uv).


Intro

Pal MCP Server is a multi-model AI gateway that lets Claude Code, Gemini CLI, and Codex CLI use multiple LLM providers as one unified system, with 11,300+ GitHub stars. It supports 7+ providers — Gemini, OpenAI, Azure OpenAI, X.AI/Grok, OpenRouter, DIAL, and Ollama (local) — with features like model routing, CLI-to-CLI bridging (clink), and conversation continuity across models. When Claude's context resets, other models can "remind" Claude of the full discussion.

Best for developers who want to leverage multiple AI models from a single Claude Code session. Works with: Claude Code, Gemini CLI, Codex CLI. Setup time: under 5 minutes.


Core Tools (Enabled by Default)

Tool Description
chat Send messages to any supported model
thinkdeep Extended reasoning with model selection
planner Multi-step project planning across models
consensus Get multiple model opinions and find consensus
codereview Cross-model code review
precommit Pre-commit checks using multiple models
debug Cross-model debugging assistance
apilookup API documentation lookup via models
challenge Devil's advocate analysis from another model
clink CLI-to-CLI bridge — spawn Codex/Gemini CLI subagents

Additional Tools (Disabled by Default)

Enable by removing from DISABLED_TOOLS env var:

Tool Description
analyze Deep code analysis
refactor Code refactoring suggestions
testgen Test generation
secaudit Security auditing
docgen Documentation generation
tracer Code flow tracing

Supported Providers

Provider Models API Key Env Var
Gemini gemini-2.5-pro, gemini-2.5-flash GEMINI_API_KEY
OpenAI gpt-4o, o3, o4-mini OPENAI_API_KEY
Azure OpenAI Any deployed model AZURE_OPENAI_*
X.AI/Grok grok-3 XAI_API_KEY
OpenRouter 200+ models OPENROUTER_API_KEY
Ollama Any local model (local, no key)

Key Feature: clink (CLI-to-CLI Bridge)

The clink tool lets Claude Code spawn Codex or Gemini CLI as isolated subagents for specific tasks — code reviews, bug hunting, research — without polluting Claude's main context window. Results flow back automatically.

Key Feature: Conversation Continuity

Full context flows across tools and models. When Claude's context resets mid-session, other models retain the conversation history and can bring Claude back up to speed.

FAQ

Q: What is Pal MCP Server? A: An MCP server that connects Claude Code to 7+ AI providers (Gemini, OpenAI, Grok, Ollama, etc.), enabling multi-model workflows, cross-model code review, and CLI-to-CLI bridging.

Q: Is Pal MCP Server free? A: The server itself is free and open source. You need API keys for the model providers you want to use (Gemini has a free tier).

Q: How do I install Pal MCP Server? A: Add the JSON config to your .mcp.json file and set your API keys. Requires Python 3.10+ and uv.


🙏

Source & Thanks

Created by BeehiveInnovations. Licensed under custom license.

pal-mcp-server — ⭐ 11,300+

Thank you for building a powerful multi-model gateway for the AI developer community.

Discussion

Sign in to join the discussion.
No comments yet. Be the first to share your thoughts.

Related Assets