{"version":"1.0","workflow_uuid":"269ce92b-4558-11f1-9bc6-00163e2b0d79","workflow_title":"Shimmy — Python-Free Rust Inference Server for Local LLMs","install_contract":{"version":"1.0","installReady":false,"title":"Shimmy — Python-Free Rust Inference Server for Local LLMs","summary":"Shimmy is a single-binary Rust inference server that serves GGUF and SafeTensors models via an OpenAI-compatible API, with hot model swapping and auto-discovery.","assetType":"Scripts","pageUrl":"https://tokrepo.com/en/workflows/asset-269ce92b","sourceUrl":"https://github.com/Michael-A-Kuykendall/shimmy/releases/latest/download/shimmy-linux-x86_64","intendedFor":[],"firstActions":[],"agentFirstSteps":[],"targetPaths":[],"verification":[],"startingPoints":[],"example":"","successOutcome":"","boundaries":[],"askUserIf":["the current workspace stack cannot be matched to a safe upstream template","the target path is not the project root, or an existing file should be merged instead of overwritten"]}}