TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

Practical Tips for Finetuning LLMs Using LoRA (Low-Rank Adaptation)

342 点作者 rasbt超过 1 年前

9 条评论

kromem超过 1 年前
I&#x27;ve been increasingly wondering if the field considering LLMs as a continuum as opposed to a set of distinct thresholds is leading to erroneous &quot;rules of thumb&quot; as most research on methodology is concentrated in smaller and more accessible model experimentation right now.<p>We generally recognize (nearly ad nauseum) that mouse models of medical research don&#x27;t necessarily translate to humans.<p>Similarly, I&#x27;d imagine most would laugh at the idea that a neurology researcher who found the best way to get a fruit fly&#x27;s brain to navigate a maze should extrapolate that methodology to a dolphin or a chimp&#x27;s brain.<p>Maybe we should be defining &quot;weight classes&quot; for LLMs and grouping research based on those classes. Like &quot;these are the techniques that work best for lightweight models&quot; but not necessarily assuming those as a general rule of thumb for &quot;heavyweight models.&quot;<p>Even something like the discussion of synthetic data on model collapse is a good example of where there might be a very significant difference in the effect on model quality for a cheaper and less sophisticated model generating synthetic data to feed back into itself and a much more complex and sophisticated model. Maybe the lesson is actually &quot;recursive training on synthetic data leads to model collapse <i>in lightweight and medium weight models</i>.&quot;<p>So while the writeup is a great one on fine tuning 7B models with LoRA, I would be curious just what % of the recommendations hold true in replication for even just a 65B model.
评论 #38341392 未加载
nl超过 1 年前
This is an exceptionally useful article. A few highlights:<p>* QLoRA works really well compared to LoRA if you need to save memory (at the cost of time)<p>* For small LoRAs, Adam has almost no memory usage penalty compared to SGD<p>* Multiple training epochs lower performance (!). To quote: &quot;This performance decline is likely due to increased overfitting, which warrants additional investigation.&quot; (Note that this is LoRA overfitting, and unclear which layers it was enabled for for this experiment).<p>* The best results for alpha and r parameters in LoRA seems to be alpha = 2r.<p>* Better datasets are much better. 1k LIMA gives better results than 50k Alpaca
评论 #38343518 未加载
hkonsti超过 1 年前
LoRA blew me away the first time I looked into it. Especially since you can host many LoRA adapters at once for a fraction of the cost of hosting an entire model by sharing the base between the adapters. I built a little tool to make LoRA fine-tuning easier. The adapters export to Huggingface. You can check it out here: <a href="https:&#x2F;&#x2F;app.haven.run">https:&#x2F;&#x2F;app.haven.run</a>
评论 #38340698 未加载
Nevin1901超过 1 年前
I fine tuned LLama-2 on code&#x2F;comment generation (in python) for around $2 and was able to run it natively on an m1 macbook air. I can totally see smaller fine tuned LLM&#x27;s being used locally on consumer devices in the future. I think people underestimate how cheap and efficient this stuff is.<p>I&#x27;ve actually built a service which lets you fine tune LLama-2&#x2F;other llms by uploading a JSON dataset. I&#x27;m looking for feedback, the link is <a href="https:&#x2F;&#x2F;useftn.com" rel="nofollow noreferrer">https:&#x2F;&#x2F;useftn.com</a>.
评论 #38342752 未加载
评论 #38342076 未加载
simonw超过 1 年前
I&#x27;m still waiting for someone to publish a &quot;LoRA in ten steps&quot; document, with working code, aimed at impatient people like me.
评论 #38344844 未加载
millerhooks超过 1 年前
I’ve been thinking about ways to compress information with ai for long distance transmission with LoRA radio for a while and now this LoRA in the news gets me all confused.
sandGorgon超过 1 年前
what is the toolset that works best ?<p>axolotl is generally recommended...but unsure if that is what is genuinely the best for production scale finetuning.
xgbzvzj超过 1 年前
शगहस ह्षब डीबीएनएस
评论 #38346737 未加载
behnamoh超过 1 年前
Ever since the author paywalled some of his useful posts, I stopped following him. I have read his ML book and I know he used to be a professor and is now working in the industry, and he’s quite famous in the field. That’s why I don’t understand why such a figure would even need the extra income generated by Substack’s paywall.
评论 #38340014 未加载
评论 #38340047 未加载