{"version":"1.0","workflow_uuid":"0f835fd8-366d-11f1-9bc6-00163e2b0d79","workflow_title":"Ollama — Run Large Language Models Locally","install_contract":{"version":"1.0","installReady":false,"title":"Ollama — Run Large Language Models Locally","summary":"Ollama makes it easy to run open-source large language models like Llama 3, Mistral, Gemma, and Qwen on your own machine. A single command downloads and runs any model with optimized inference, GPU acceleration, and a REST API.","assetType":"Scripts","pageUrl":"https://tokrepo.com/en/workflows/0f835fd8-366d-11f1-9bc6-00163e2b0d79","sourceUrl":"https://ollama.com/install.sh","intendedFor":[],"firstActions":[],"agentFirstSteps":[],"targetPaths":[],"verification":[],"startingPoints":[],"example":"","successOutcome":"","boundaries":[],"askUserIf":["the current workspace stack cannot be matched to a safe upstream template","the target path is not the project root, or an existing file should be merged instead of overwritten"]}}