{"version":"1.0","workflow_uuid":"f0ec1009-4771-11f1-9bc6-00163e2b0d79","workflow_title":"KoboldCpp — Single-File Local LLM Inference Engine","install_contract":{"version":"1.0","installReady":false,"title":"KoboldCpp — Single-File Local LLM Inference Engine","summary":"KoboldCpp is a self-contained local LLM inference engine that runs GGUF models with GPU acceleration on consumer hardware, providing an OpenAI-compatible API and built-in web UI without requiring Python or complex setup.","assetType":"Scripts","pageUrl":"https://tokrepo.com/en/workflows/asset-f0ec1009","sourceUrl":"http://localhost:5001","intendedFor":[],"firstActions":[],"agentFirstSteps":[],"targetPaths":[],"verification":[],"startingPoints":[],"example":"","successOutcome":"","boundaries":[],"askUserIf":["the current workspace stack cannot be matched to a safe upstream template","the target path is not the project root, or an existing file should be merged instead of overwritten"]}}