{"version":"1.0","workflow_uuid":"bc541841-470d-11f1-9bc6-00163e2b0d79","workflow_title":"LMCache — Supercharge LLM Inference with KV Cache Sharing","install_contract":{"version":"1.0","installReady":false,"title":"LMCache — Supercharge LLM Inference with KV Cache Sharing","summary":"LMCache is an open-source KV cache management layer that accelerates LLM inference by sharing and reusing key-value caches across requests, reducing time-to-first-token and GPU memory usage.","assetType":"Configs","pageUrl":"https://tokrepo.com/en/workflows/asset-bc541841","sourceUrl":"https://github.com/LMCache/LMCache","intendedFor":[],"firstActions":[],"agentFirstSteps":[],"targetPaths":[],"verification":[],"startingPoints":[],"example":"","successOutcome":"","boundaries":[],"askUserIf":["the current workspace stack cannot be matched to a safe upstream template","the target path is not the project root, or an existing file should be merged instead of overwritten"]}}