Cette page est affichée en anglais. Une traduction française est en cours.
WorkflowsMay 12, 2026·2 min de lecture

RagaAI Catalyst — LLM Eval + Tracing SDK

RagaAI Catalyst is a Python SDK for managing LLM projects with evaluation, dataset management, trace/agentic tracing, and prompt/guardrail workflows.

Prêt pour agents

Cet actif peut être lu et installé directement par les agents

TokRepo expose une commande CLI universelle, un contrat d'installation, le metadata JSON, un plan selon l'adaptateur et le contenu raw pour aider les agents à juger l'adaptation, le risque et les prochaines actions.

Native · 94/100Policy : autoriser
Surface agent
Tout agent MCP/CLI
Type
Cli
Installation
Manual
Confiance
Confiance : Established
Point d'entrée
pip install ragaai-catalyst
Commande CLI universelle
npx tokrepo install 4c25e454-4724-5d35-942e-50bdbcbc1b86
Introduction

RagaAI Catalyst is a Python SDK for managing LLM projects with evaluation, dataset management, trace/agentic tracing, and prompt/guardrail workflows.

  • Best for: Teams that need repeatable evals, tracing, and guardrails for production LLM apps
  • Works with: Python; your Catalyst credentials (access/secret keys) per README; integrates with LLM pipelines
  • Setup time: 15–45 minutes

Practical Notes

  • GitHub: 16,156 stars · 2,019 forks; pushed 2026-02-11 (verified via GitHub API).
  • README installation is pip install ragaai-catalyst and config uses access_key / secret_key / base_url.
  • README lists modules for evaluation, trace management, agentic tracing, prompt management, and guardrails.

Main

A practical way to adopt evaluation:

  1. Define a “golden set” of prompts + expected behaviors, and keep it versioned.
  2. Instrument tracing first, so every regression can be tied to a specific change (prompt/model/tooling).
  3. Treat guardrails as tests: start with allowlists/denylists, then add heuristic checks and human review gates.
  4. Track cost and latency next to quality; a “better” model that doubles latency may not be viable.

Make evals run on every release candidate, not just ad-hoc experiments.

FAQ

Q: Is it only for evaluation? A: No—README includes tracing, prompt management, and guardrail/red-teaming modules too.

Q: Do I need credentials? A: Yes—README config uses access and secret keys plus a base URL before operations.

Q: What should I measure first? A: Start with correctness and safety, then add latency and cost as first-class metrics.

🙏

Source et remerciements

Source: https://github.com/raga-ai-hub/RagaAI-Catalyst > License: Apache-2.0 > GitHub stars: 16,156 · forks: 3,607

Fil de discussion

Connectez-vous pour rejoindre la discussion.
Aucun commentaire pour l'instant. Soyez le premier à partager votre avis.

Actifs similaires