{"version":"1.0","workflow_uuid":"98b30827-ac0a-4067-8f64-7360f2a13995","workflow_title":"Ollama — Run LLMs Locally with One Command","install_contract":{"version":"1.0","installReady":false,"title":"Ollama — Run LLMs Locally with One Command","summary":"Run Llama 3, Mistral, Gemma, Phi, and 100+ open-source LLMs locally with a single command. OpenAI-compatible API for seamless integration with AI tools. 120,000+ GitHub stars.","assetType":"Configs","pageUrl":"https://tokrepo.com/en/workflows/98b30827-ac0a-4067-8f64-7360f2a13995","sourceUrl":"https://github.com/ollama","intendedFor":[],"firstActions":[],"agentFirstSteps":[],"targetPaths":[],"verification":[],"startingPoints":[],"example":"","successOutcome":"","boundaries":[],"askUserIf":["the current workspace stack cannot be matched to a safe upstream template","the target path is not the project root, or an existing file should be merged instead of overwritten"]}}