{"version":"1.0","workflow_uuid":"b1962c77-9ecf-4a84-87b1-e7d4b677dabe","workflow_title":"Cloudflare AI Gateway — LLM Proxy, Cache & Analytics","install_contract":{"version":"1.0","installReady":false,"title":"Cloudflare AI Gateway — LLM Proxy, Cache & Analytics","summary":"Free proxy gateway for LLM API calls with caching, rate limiting, cost tracking, and fallback routing across providers. Reduce costs up to 95% with response caching. 7,000+ stars.","assetType":"Configs","pageUrl":"https://tokrepo.com/en/workflows/b1962c77-9ecf-4a84-87b1-e7d4b677dabe","sourceUrl":"https://github.com/cloudflare","intendedFor":[],"firstActions":[],"agentFirstSteps":[],"targetPaths":[],"verification":[],"startingPoints":[],"example":"","successOutcome":"","boundaries":[],"askUserIf":["the current workspace stack cannot be matched to a safe upstream template","the target path is not the project root, or an existing file should be merged instead of overwritten"]}}