TE
TechEcho
Home
24h Top
Newest
Best
Ask
Show
Jobs
English
GitHub
Twitter
Home
QLoRA 4-bit finetuning of LLMs
7 points
by
kashifr
almost 2 years ago
1 comment
kashifr
almost 2 years ago
An efficient finetuning approach that reduces memory usage enough to finetune a 65B parameter model on a single 48GB GPU while preserving full 16-bit finetuning task performance!