{"version":"1.0","workflow_uuid":"487e41a3-6e23-4d5b-97c3-57c2ed5c6c87","workflow_title":"Claude Code Agent: Model Evaluator — Benchmark AI Models","install_contract":{"version":"1.0","installReady":false,"title":"Claude Code Agent: Model Evaluator — Benchmark AI Models","summary":"Claude Code agent for evaluating and benchmarking LLM outputs. Compare models, measure quality, and track performance metrics.","assetType":"Skills","pageUrl":"https://tokrepo.com/en/workflows/claude-code-agent-model-evaluator-benchmark-ai-models-487e41a3","sourceUrl":"https://tokrepo.com/en/workflows/1cf2f5bc-ce0e-4242-ab2f-34ad488b478e","intendedFor":[],"firstActions":[],"agentFirstSteps":[],"targetPaths":[],"verification":[],"startingPoints":[],"example":"","successOutcome":"","boundaries":[],"askUserIf":["the current workspace stack cannot be matched to a safe upstream template","the target path is not the project root, or an existing file should be merged instead of overwritten"]}}