{"version":"1.0","workflow_uuid":"08f46b1e-5a82-59dc-916a-cb2f0ae17a63","workflow_title":"Giskard Checks — Evals and Safety Tests for LLM Agents","install_contract":{"version":"1.0","installReady":false,"title":"Giskard Checks — Evals and Safety Tests for LLM Agents","summary":"Giskard Checks gives Python teams a modular eval layer for agent regressions, groundedness, and policy conformance with scenario-based tests.","assetType":"Scripts","pageUrl":"","sourceUrl":"https://github.com/Giskard-AI/giskard-oss","intendedFor":[],"firstActions":[],"agentFirstSteps":[],"targetPaths":[],"verification":[],"startingPoints":[],"example":"","successOutcome":"","boundaries":[],"askUserIf":["the current workspace stack cannot be matched to a safe upstream template","the target path is not the project root, or an existing file should be merged instead of overwritten"]}}