{"version":"1.0","workflow_uuid":"d1888a22-7087-4310-bcaa-dca6663a2e18","workflow_title":"llm-guard — Secure LLM Inputs & Outputs","install_contract":{"version":"1.0","installReady":false,"title":"llm-guard — Secure LLM Inputs & Outputs","summary":"Harden LLM apps with a scanner pipeline for prompt injection, PII leakage, toxicity, and unsafe output. Install in minutes and gate requests in code.","assetType":"Scripts","pageUrl":"","sourceUrl":"https://github.com/protectai/llm-guard","intendedFor":[],"firstActions":[],"agentFirstSteps":[],"targetPaths":[],"verification":[],"startingPoints":[],"example":"","successOutcome":"","boundaries":[],"askUserIf":["the current workspace stack cannot be matched to a safe upstream template","the target path is not the project root, or an existing file should be merged instead of overwritten"]}}