{"version":"1.0","workflow_uuid":"6469e991-3d9d-11f1-9bc6-00163e2b0d79","workflow_title":"WebLLM — Run Large Language Models Directly in the Browser","install_contract":{"version":"1.0","installReady":false,"title":"WebLLM — Run Large Language Models Directly in the Browser","summary":"WebLLM is an MLC project that brings LLM inference to web browsers using WebGPU. It runs models like LLaMA, Mistral, and Phi entirely client-side with no server required, enabling private AI chat and text generation from any modern browser.","assetType":"Scripts","pageUrl":"https://tokrepo.com/en/workflows/6469e991-3d9d-11f1-9bc6-00163e2b0d79","sourceUrl":"https://github.com/mlc-ai/web-llm","intendedFor":[],"firstActions":[],"agentFirstSteps":[],"targetPaths":[],"verification":[],"startingPoints":[],"example":"","successOutcome":"","boundaries":[],"askUserIf":["the current workspace stack cannot be matched to a safe upstream template","the target path is not the project root, or an existing file should be merged instead of overwritten"]}}