{"version":"1.0","workflow_uuid":"f10382e2-4fa6-491f-8110-84f8397af129","workflow_title":"Guardrails — Validate & Secure LLM Outputs","install_contract":{"version":"1.0","installReady":false,"title":"Guardrails — Validate & Secure LLM Outputs","summary":"Guardrails is a Python framework for validating LLM inputs/outputs to detect risks and generate structured data. 6.6K+ GitHub stars. Pre-built validators, Pydantic models. Apache 2.0.","assetType":"Scripts","pageUrl":"https://tokrepo.com/en/workflows/guardrails-validate-secure-llm-outputs-f10382e2","sourceUrl":"https://github.com/guardrails-ai","intendedFor":[],"firstActions":[],"agentFirstSteps":[],"targetPaths":[],"verification":[],"startingPoints":[],"example":"","successOutcome":"","boundaries":[],"askUserIf":["the current workspace stack cannot be matched to a safe upstream template","the target path is not the project root, or an existing file should be merged instead of overwritten"]}}