TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

How to use Alpaca-LoRA to fine-tune a model like ChatGPT

173 点作者 bfirsh大约 2 年前

13 条评论

rishsriv大约 2 年前
This looks fantastic. Will try replacing our current fine-tuned FLAN-UL2 model with this.<p>I wonder how the devtooling around this will evolve. Seems like a matter of days until someone creates a GUI wrapper around this, and obviates the need to use programmer time for fine-tuning
评论 #35282481 未加载
isoprophlex大约 2 年前
<i>Low-rank adaptation (LoRA) ... has some advantages over previous methods:</i><p>- <i>It is faster and uses less memory, which means it can run on consumer hardware.</i><p>- <i>The output is much smaller (megabytes, not gigabytes).</i><p>- <i>You can combine multiple fine-tuned models together at runtime.</i><p>This is great news for my dream of building a fine-tuned interactive messenger, that can deliver a message on my behalf by training it on my personality &amp; the information I want to convey.<p>Now just add text to speech and a talking head, as discussed in that other submission about cloning yourself with AI... <a href="https:&#x2F;&#x2F;news.ycombinator.com&#x2F;item?id=35280418" rel="nofollow">https:&#x2F;&#x2F;news.ycombinator.com&#x2F;item?id=35280418</a>
评论 #35281375 未加载
评论 #35281354 未加载
评论 #35281149 未加载
camdenlock大约 2 年前
&gt; The weights for LLaMA have not yet been released publicly. To apply for access, fill out this Meta Research form.<p>Cute. ;)
评论 #35280811 未加载
tysam_and大约 2 年前
LoRA has actually been around for a little while! I first saw it when it became popular in fine-tuning models quantized down to about 8 bits or so. I&#x27;m sure it&#x27;s doing stuff in the 4bit range now! :D<p>I believe it&#x27;s a core toolbox piece of tech required to really push the limits of LLMs either in original training or in inference. Similar sort of to how batch norm was for convolutional neural networks. I look forward to seeing how this will be applied in the future.
syntaxing大约 2 年前
The easiest way to run alpaca Lora locally is with this little known fork [1] that uses Docker. You’ll be up and running in about 20 min with pretty much any modern consumer Nvidia GPU.<p>[1] <a href="https:&#x2F;&#x2F;github.com&#x2F;chris-alexiuk&#x2F;alpaca-lora">https:&#x2F;&#x2F;github.com&#x2F;chris-alexiuk&#x2F;alpaca-lora</a>
mnreef大约 2 年前
Hi All, I have a noob question. I have been reading about Alpaca and Alpaca Lora. I have a use case in which I want to fine tune&#x2F;train Alpaca Lora on a large corpus of books which are in the txt format. I know for Alpaca, the data was in &quot;Instruction : Prompt&quot; format. however, my text is huge and is not in that format. It&#x27;s simply a library of books and journal articles. I want to be able to ask a question and the model answers based on the books I trained it on. I also want to be able to ask general questions for example which books discussed topic x or y.<p>I have tried OpenAI&#x27;s API to create embeddings, but I want to use Alpaca.<p>I really appreciate your help.
braingenious大约 2 年前
I love these idea of LoRAs for LLMs.<p>Has anybody made a llama&#x2F;alpaca erebus model? I read about them in the oobabooga docs and a locally-run language model fine tuned on literotica could be the funniest thing I’ve ever seen.
credit_guy大约 2 年前
I guess this LoRA is the missing piece.<p>NVIDIA stated recently that GPT bots will become one million times more powerful in ten years. Many people doubted that.<p>With LoRA, I see a much higher improvement. These guys claim a 10000 times reduction in parameter size. A different way to look at it, is that with the current hardware you can train a model that has 10000 times more parameters. If you add a 100x improvement in hardware in 10 years (not at all unrealistic), that&#x27;s the million. But we will have significant improvements in training methods too.
评论 #35281766 未加载
nico大约 2 年前
Can a model be fine-tuned “online”?<p>If cost wasn’t an issue, could I fine-tune a model in real time, while also using it for inference?
评论 #35283151 未加载
评论 #35284038 未加载
评论 #35283401 未加载
rcarmo大约 2 年前
So they use cog before installing it? Apparently this wasn’t proofread.<p>Also, is it just me or there are currently more ways to run LLMs on a CPU than on a GPU springing up on GitHub? I have hacked my own, but my chat UI is awful, so what is the nicest, pre-packaged CUDA-friendly way to run this now?
eachro大约 2 年前
How does LoRA save more than 50% of the memory usage? I see that the weight updates have much lower memory footprint by virtue if being low rank. But you still need the dense weights for the forward pass dont you?
评论 #35284326 未加载
slicktux大约 2 年前
Anyone else click on this thinking it was about the wireless protocol?
评论 #35282228 未加载
techn00大约 2 年前
It feels like I&#x27;m living in a cartoon with all these terms: &gt; In this blog post, we’ll show you how to use LoRA to fine-tune LLaMA using Alpaca training data.