{"version":"1.0","workflow_uuid":"fa6e0b07-4c49-11f1-9bc6-00163e2b0d79","workflow_title":"Liger-Kernel — Efficient GPU Kernels for LLM Training","install_contract":{"version":"1.0","installReady":false,"title":"Liger-Kernel — Efficient GPU Kernels for LLM Training","summary":"Liger-Kernel provides optimized Triton kernels for LLM training that reduce GPU memory usage and improve throughput, serving as drop-in replacements for standard HuggingFace Transformers layers.","assetType":"Configs","pageUrl":"https://tokrepo.com/en/workflows/asset-fa6e0b07","sourceUrl":"https://github.com/linkedin/Liger-Kernel","intendedFor":[],"firstActions":[],"agentFirstSteps":[],"targetPaths":[],"verification":[],"startingPoints":[],"example":"","successOutcome":"","boundaries":[],"askUserIf":["the current workspace stack cannot be matched to a safe upstream template","the target path is not the project root, or an existing file should be merged instead of overwritten"]}}