{"version":"1.0","workflow_uuid":"ffbad589-cd32-4eca-9518-fdcf9167ca21","workflow_title":"Guardrails AI — Validate LLM Outputs in Production","install_contract":{"version":"1.0","installReady":false,"title":"Guardrails AI — Validate LLM Outputs in Production","summary":"Add validation and guardrails to any LLM output. Guardrails AI checks for hallucination, toxicity, PII leakage, and format compliance with 50+ built-in validators.","assetType":"Workflows","pageUrl":"https://tokrepo.com/en/workflows/ffbad589-cd32-4eca-9518-fdcf9167ca21","sourceUrl":"https://github.com/guardrails-ai","intendedFor":[],"firstActions":[],"agentFirstSteps":[],"targetPaths":[],"verification":[],"startingPoints":[],"example":"","successOutcome":"","boundaries":[],"askUserIf":["the current workspace stack cannot be matched to a safe upstream template","the target path is not the project root, or an existing file should be merged instead of overwritten"]}}