TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

LoRA+: Efficient Low Rank Adaptation of Large Models

181 点作者 veryluckyxyz大约 1 年前

7 条评论

batterseapower大约 1 年前
The other recent improvement suggested for LoRA is DoRA: <a href="https:&#x2F;&#x2F;magazine.sebastianraschka.com&#x2F;p&#x2F;lora-and-dora-from-scratch" rel="nofollow">https:&#x2F;&#x2F;magazine.sebastianraschka.com&#x2F;p&#x2F;lora-and-dora-from-s...</a>. It really does seem to strongly outperform LoRA - see also <a href="https:&#x2F;&#x2F;www.answer.ai&#x2F;posts&#x2F;2024-04-26-fsdp-qdora-llama3.html" rel="nofollow">https:&#x2F;&#x2F;www.answer.ai&#x2F;posts&#x2F;2024-04-26-fsdp-qdora-llama3.htm...</a>
评论 #40189499 未加载
评论 #40189472 未加载
评论 #40189647 未加载
评论 #40192619 未加载
cuuupid大约 1 年前
I’m struggling to understand from this paper whether the approach is better in the general sense (all cases, with wider models seeing greater benefits) or purely for wider models (with narrower models seeing detriment)?<p>If it’s the former this could effectively halve finetuning cost overnight which would go a significant way towards enabling a wider array of use cases for LoRA.
ironbound大约 1 年前
I&#x27;ve had sucess with GaLore: Memory-Efficient LLM Training by Gradient Low-Rank Projection <a href="https:&#x2F;&#x2F;arxiv.org&#x2F;abs&#x2F;2403.03507" rel="nofollow">https:&#x2F;&#x2F;arxiv.org&#x2F;abs&#x2F;2403.03507</a>
评论 #40194836 未加载
youssefabdelm大约 1 年前
A better name would&#x27;ve probably been FastLoRA or something
评论 #40191477 未加载
评论 #40189791 未加载
yau8edq12i大约 1 年前
What an unfortunate name... I initially thought this was about wireless communication. <a href="https:&#x2F;&#x2F;en.wikipedia.org&#x2F;wiki&#x2F;LoRa" rel="nofollow">https:&#x2F;&#x2F;en.wikipedia.org&#x2F;wiki&#x2F;LoRa</a>
评论 #40189098 未加载
评论 #40189258 未加载
评论 #40189190 未加载
评论 #40188920 未加载
评论 #40189002 未加载
allpaca大约 1 年前
This is old, having been released in February... Why do you talk about it now?
axpy906大约 1 年前
In 2024 are folks still swapping out LoRA adapters? Is this still relevant?
评论 #40189366 未加载
评论 #40189356 未加载