{"version":"1.0","workflow_uuid":"4a86c01b-d4d2-4c0f-8152-393c5685e429","workflow_title":"Instructor — Structured Outputs from LLMs","install_contract":{"version":"1.0","installReady":false,"title":"Instructor — Structured Outputs from LLMs","summary":"Get structured, validated outputs from LLMs using Pydantic models. Works with OpenAI, Anthropic, Google, Ollama, and more. Retry logic, streaming, partial responses. 12.6K+ stars.","assetType":"Scripts","pageUrl":"https://tokrepo.com/en/workflows/instructor-structured-outputs-llms-4a86c01b","sourceUrl":"https://github.com/instructor-ai","intendedFor":[],"firstActions":[],"agentFirstSteps":[],"targetPaths":[],"verification":[],"startingPoints":[],"example":"","successOutcome":"","boundaries":[],"askUserIf":["the current workspace stack cannot be matched to a safe upstream template","the target path is not the project root, or an existing file should be merged instead of overwritten"]}}