{"version":"1.0","workflow_uuid":"ee93312c-3d9c-11f1-9bc6-00163e2b0d79","workflow_title":"PEFT — Parameter-Efficient Fine-Tuning for Large Language Models","install_contract":{"version":"1.0","installReady":false,"title":"PEFT — Parameter-Efficient Fine-Tuning for Large Language Models","summary":"PEFT is a Hugging Face library for adapting large pre-trained models using parameter-efficient methods like LoRA, QLoRA, prompt tuning, and prefix tuning. It enables fine-tuning billion-parameter models on consumer hardware by updating only a small fraction of weights.","assetType":"Configs","pageUrl":"https://tokrepo.com/en/workflows/ee93312c-3d9c-11f1-9bc6-00163e2b0d79","sourceUrl":"https://github.com/huggingface/peft","intendedFor":[],"firstActions":[],"agentFirstSteps":[],"targetPaths":[],"verification":[],"startingPoints":[],"example":"","successOutcome":"","boundaries":[],"askUserIf":["the current workspace stack cannot be matched to a safe upstream template","the target path is not the project root, or an existing file should be merged instead of overwritten"]}}