{"version":"1.0","workflow_uuid":"6fa2ec3f-42dc-11f1-9bc6-00163e2b0d79","workflow_title":"bitsandbytes — Accessible Large Language Model Quantization","install_contract":{"version":"1.0","installReady":false,"title":"bitsandbytes — Accessible Large Language Model Quantization","summary":"Lightweight CUDA library for 8-bit and 4-bit quantization, enabling large model fine-tuning and inference on consumer GPUs with minimal accuracy loss.","assetType":"Scripts","pageUrl":"https://tokrepo.com/en/workflows/6fa2ec3f-42dc-11f1-9bc6-00163e2b0d79","sourceUrl":"https://github.com/bitsandbytes-foundation/bitsandbytes","intendedFor":[],"firstActions":[],"agentFirstSteps":[],"targetPaths":[],"verification":[],"startingPoints":[],"example":"","successOutcome":"","boundaries":[],"askUserIf":["the current workspace stack cannot be matched to a safe upstream template","the target path is not the project root, or an existing file should be merged instead of overwritten"]}}