SkillsApr 8, 2026·2 min read

Together AI GPU Clusters Skill for Claude Code

Skill that teaches Claude Code Together AI's GPU cluster API. Provision on-demand and reserved H100, H200, and B200 GPU clusters for large-scale training and inference.

TL;DR
A Claude Code skill that teaches the agent how to provision and manage H100, H200, and B200 GPU clusters on Together AI.
§01

What it is

The Together AI GPU Clusters skill teaches Claude Code how to use Together AI's GPU cluster API. Once installed, Claude Code can provision on-demand and reserved clusters of H100, H200, and B200 GPUs for large-scale model training and distributed inference. The skill provides the agent with API knowledge, parameter schemas, and best practices for cluster management.

ML engineers running large training jobs, teams needing burst GPU capacity, and organizations evaluating GPU cloud providers will find this skill useful for managing Together AI infrastructure directly from their development environment.

§02

How it saves time or tokens

Managing GPU clusters typically requires navigating cloud dashboards, writing infrastructure scripts, and reading API documentation. This skill embeds the Together AI GPU API knowledge directly into Claude Code, so you can describe your compute needs in natural language and the agent generates the correct API calls. It handles cluster sizing, pricing tier selection, and configuration details without manual documentation lookup.

§03

How to use

  1. Install the skill:
npx skills add togethercomputer/skills
  1. Start Claude Code in your project.
  2. Ask the agent to provision GPU resources:
> Provision an 8xH100 cluster for a 70B model training run
> List my active GPU clusters and their costs
> Reserve 4xB200 GPUs for next month
§04

Example

# Install the skill
npx skills add togethercomputer/skills

# In Claude Code session:
> I need to fine-tune a 13B parameter model.
  What GPU cluster configuration do you recommend on Together AI?

# The agent will:
# - Recommend cluster size based on model parameters
# - Show pricing for on-demand vs reserved options
# - Generate the API call to provision the cluster
# - Explain monitoring and teardown procedures
§05

Related on TokRepo

§06

Common pitfalls

  • GPU clusters incur costs immediately upon provisioning. Always confirm pricing and set auto-shutdown policies before launching clusters.
  • Reserved instances require commitment periods. Understand the reservation terms before switching from on-demand to reserved pricing.
  • The skill teaches Claude Code the API but does not manage billing. Monitor your Together AI dashboard for usage and cost tracking.

Frequently Asked Questions

What GPU types are available through Together AI?+

Together AI offers H100, H200, and B200 GPU clusters. These are available in various configurations for on-demand and reserved usage. The skill helps Claude Code recommend the right GPU type and cluster size based on your workload requirements.

How do I install the Together AI GPU skill?+

Run npx skills add togethercomputer/skills in your terminal. This installs the skill files into your Claude Code configuration. The agent automatically uses the skill knowledge when GPU provisioning tasks are detected.

Can the skill help estimate training costs?+

Yes. The skill includes pricing information for different GPU types and commitment levels. Claude Code can estimate costs based on your training duration, cluster size, and whether you use on-demand or reserved instances.

Does the skill support inference workloads?+

Yes. The Together AI GPU cluster API supports both training and inference workloads. The skill covers provisioning clusters for distributed inference, including configuration for model serving and endpoint management.

Do I need a Together AI account to use this skill?+

Yes. You need a Together AI account with API access and appropriate credits or billing configured. The skill provides the API knowledge to Claude Code, but actual provisioning requires valid Together AI credentials.

Citations (3)
🙏

Source & Thanks

Part of togethercomputer/skills — MIT licensed.

Discussion

Sign in to join the discussion.
No comments yet. Be the first to share your thoughts.

Related Assets