{"version":"1.0","workflow_uuid":"0eefb7ad-754e-4f35-8967-586ebf4c2a6a","workflow_title":"Ollama — Run LLMs Locally","install_contract":{"version":"1.0","installReady":false,"title":"Ollama — Run LLMs Locally","summary":"Run large language models locally on your machine. Supports Llama 3, Mistral, Gemma, Phi, and dozens more. One-command install, OpenAI-compatible API.","assetType":"Scripts","pageUrl":"https://tokrepo.com/en/workflows/ollama-run-llms-locally-0eefb7ad","sourceUrl":"https://ollama.com/install.sh","intendedFor":[],"firstActions":[],"agentFirstSteps":[],"targetPaths":[],"verification":[],"startingPoints":[],"example":"","successOutcome":"","boundaries":[],"askUserIf":["the current workspace stack cannot be matched to a safe upstream template","the target path is not the project root, or an existing file should be merged instead of overwritten"]}}