TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

The genie escapes: Stanford copies the ChatGPT AI for less than $600

207 点作者 Freddie111大约 2 年前

23 条评论

superkuh大约 2 年前
Hardly. I&#x27;ve played a lot with the 7,13, and 30B llamas as well as the 7 and 13B alpacas fine tuned by Stanford. They do not have emergent abilities like being able to generate rhymes or, say, represent a movie plot as emoji. Even openai&#x27;s old text-davinci-003 (gpt3.5, but text completion, not the chat ones) far outperforms them. That said, I have hopes for a 65B 3-bit quantized alpaca-fine tuned. We&#x27;ll see when someone spends the money to do the (more costly) 65B training. The alpacas are also much more likely to go off rails and start regurgitating their fine-tuning inputs. Either that or openai is doing a <i>lot</i> of post processing on their end to hide the same problems in their LLM.<p>For now my IRC bots run the alpaca 7B 4-bit. 13B was not a significant improvement for twice the computational time. But it&#x27;s best to learn them now because as soon as openai gets sued for the first time all the turing test passing older models without the legal-butt-covering bolted on will be removed.
评论 #35238884 未加载
评论 #35240051 未加载
评论 #35239737 未加载
评论 #35239796 未加载
评论 #35238871 未加载
评论 #35239794 未加载
评论 #35238915 未加载
评论 #35240007 未加载
评论 #35238846 未加载
评论 #35239848 未加载
评论 #35239503 未加载
评论 #35238904 未加载
freediver大约 2 年前
The incredible contribution of Alpaca is showing the world how to efficicently train LLM on instructions. The fact that it did so on 52k instructions generated by GPT is poetic.<p>It does not matter what current capabilities of open source models are, because this opens the door to tremendous democratization of the ability to train and self-deploy these models.<p>In less than 6 months we will have open source models with gpt3-like capabilities, running locally on laptops, and potentially in phones and web browsers.
评论 #35239156 未加载
评论 #35322549 未加载
doctoboggan大约 2 年前
I&#x27;ve used both the 7B and 13B instruction tuned llama weights (quantized using the llama.cpp scripts). Either I am doing something wrong, or these two models are no-where near the level of ChatGPT. Many times they return something totally irrelevant to my question, stop responding, use a different language, or otherwise return the wrong answer. ChatGPT does none of this. (other than the wrong answer due to hallucinating sometimes...)<p>Reading through the README and issues on the llama.cpp project, there is some speculation that there is a bug in the quantization, or possibly a bug in the inference (less likely I think).<p>I hope this is true and once fixed the models can perform up to or past the ChatGPT level. If its not true and these models are performing correctly, then either the metrics used to compare it to GPT is garbage and don&#x27;t capture the real world uses, or the instruction tuning done by the Stanford team is not up to par.
评论 #35239454 未加载
Waterluvian大约 2 年前
If you use consciousness as a baseline, the intellectual difference between a grade schooler and a PhD is tiny.<p>This is what I think comparing these bots is like. You can argue that they’re very close. But the delta makes a very big difference for any practical purposes because we’re looking for nuanced capability.
评论 #35241322 未加载
braingenious大约 2 年前
I have not found alpaca to be comparable to chatgpt, but it could be because of bugs in the version I installed through dalai. I might try reinstalling it because I suspect there might be some sort of file corruption issue or whatever.<p>I gave it the prompt “cats aren’t always fuzzy” and it wrote a lengthy livejournal-esque rambling journal entry about a woman and her husband having money issues. It was funny, but lightyears away from chatgpt.<p>It does sometimes create some really funny hallucinations though, like inventing prefectures in Japan that don’t exist etc.
评论 #35239280 未加载
awinter-py大约 2 年前
&gt; asked GPT to take 175 human-written instruction&#x2F;output pairs, and start generating more in the same style and format ... through one of OpenAI&#x27;s helpfully provided APIs, and ... the team had some 52,000 sample conversations to use in post-training the LLaMA model<p>hmm I wonder if this is essentially a probe[1] technique + relies on chatgpt already having been extensively trained<p>like did they basically exfiltrate the weights<p>1. probing per <a href="https:&#x2F;&#x2F;arxiv.org&#x2F;abs&#x2F;2102.12452" rel="nofollow">https:&#x2F;&#x2F;arxiv.org&#x2F;abs&#x2F;2102.12452</a>
gaogao大约 2 年前
Has anyone tried this yet on the 65B version? I&#x27;m curious if it knows how to rhyme and other emergent behavior, as alpace-7B does not.
评论 #35248357 未加载
dang大约 2 年前
Recent and related:<p><i>Stanford Alpaca web demo suspended “until further notice”</i> - <a href="https:&#x2F;&#x2F;news.ycombinator.com&#x2F;item?id=35200557" rel="nofollow">https:&#x2F;&#x2F;news.ycombinator.com&#x2F;item?id=35200557</a> - March 2023 (77 comments)<p><i>Stanford Alpaca, and the acceleration of on-device LLM development</i> - <a href="https:&#x2F;&#x2F;news.ycombinator.com&#x2F;item?id=35141531" rel="nofollow">https:&#x2F;&#x2F;news.ycombinator.com&#x2F;item?id=35141531</a> - March 2023 (66 comments)<p><i>Alpaca: An Instruct Tuned LLaMA 7B – Responses on par with txt-DaVinci-3</i> - <a href="https:&#x2F;&#x2F;news.ycombinator.com&#x2F;item?id=35139450" rel="nofollow">https:&#x2F;&#x2F;news.ycombinator.com&#x2F;item?id=35139450</a> - March 2023 (11 comments)<p><i>Alpaca: A strong open-source instruction-following model</i> - <a href="https:&#x2F;&#x2F;news.ycombinator.com&#x2F;item?id=35136624" rel="nofollow">https:&#x2F;&#x2F;news.ycombinator.com&#x2F;item?id=35136624</a> - March 2023 (296 comments)
simonw大约 2 年前
Related, my post &quot;Could you train a ChatGPT-beating model for $85,000 and run it in a browser?&quot; <a href="https:&#x2F;&#x2F;simonwillison.net&#x2F;2023&#x2F;Mar&#x2F;17&#x2F;beat-chatgpt-in-a-browser&#x2F;" rel="nofollow">https:&#x2F;&#x2F;simonwillison.net&#x2F;2023&#x2F;Mar&#x2F;17&#x2F;beat-chatgpt-in-a-brow...</a><p>I think you can train LLaMA 7B (the model underlying Alpaca) for around $82,000, based on the Meta Research paper about it. Then you can fine-tune it ala Alpaca for a few hundred dollars more.<p>My wilder speculation is that, if you can shrink the model down to 4GB with llama.cpp 4bit quantization, it may be possible to run it entirely in the browser (ala Stable Diffusion from the other day).
satvikpendem大约 2 年前
Alpaca is cool but it&#x27;s also not technically allowed by OpenAI&#x27;s TOS, and LLaMA is certainly not allowed to be used for non-commercial purposes. With that in mind, OpenAssistant is an Apache 2.0 licensed fully open source alternative that&#x27;s pretty good (the model is OpenAssistant&#x2F;oasst-sft-1-pythia-12b): <a href="https:&#x2F;&#x2F;huggingface.co&#x2F;spaces&#x2F;olivierdehaene&#x2F;chat-llm-streaming" rel="nofollow">https:&#x2F;&#x2F;huggingface.co&#x2F;spaces&#x2F;olivierdehaene&#x2F;chat-llm-stream...</a>.<p>I&#x27;ve found OA to be better than Alpaca but I&#x27;ll wait until the 65B 3-bit quantization efforts for Alpaca are underway to compare them.
评论 #35241790 未加载
neilellis大约 2 年前
Wow, Stanford&#x27;s Alpaca AI project is a real game-changer. The fact that it performs on par with ChatGPT but costs less than $600 to build is both exciting and terrifying. Sure, it&#x27;s great to see AI becoming more accessible, but it&#x27;s also a massive wakeup call for the potential misuse of these technologies.<p>We&#x27;ve got big names like OpenAI, Google, Apple, Meta, Baidu, and Amazon putting in serious time and money to ensure their language models are safe and ethical. However, now that we know it&#x27;s possible to build powerful AI models on a budget, it&#x27;s crucial to think about what this means for the future of AI regulation and safety.<p>This Alpaca AI project is a stark reminder that we need to have a serious conversation about the possible repercussions of AI proliferation. We can&#x27;t just sit back and assume the big companies will take care of everything. The genie is out of the bottle, and it&#x27;s time for everyone in the tech community to face the music and take responsibility for the AI revolution.
raydiatian大约 2 年前
&gt; It seems these godlike AIs are already frighteningly cheap and easy to replicate.<p>Who writes this shit?
评论 #35239021 未加载
评论 #35238882 未加载
cjohnson318大约 2 年前
&gt; It seems these godlike AIs are already frighteningly cheap and easy to replicate.<p>&quot;godlike&quot;? Really? I&#x27;m not religious, but this seems like an overreaction for something that has no agency.
评论 #35239660 未加载
评论 #35240657 未加载
评论 #35239650 未加载
welly34h大约 2 年前
Code can be abstracted into a simpler code model and deterministically recreate the old code model.<p>OpenAI is an eventually to be obsoleted initial brute force approach that will be abstracted over and over into a simpler code implementation with rules to recreate the old state.<p>kkrieger is a simple example of a tiny data model that can be deterministically rehydrated. It’s not unrealistic for AI models to become <i>a</i> seed value for a normalized code base to deterministically unpack into necessary electron state
starik36大约 2 年前
From the article: Pre-trained on a trillion &quot;tokens&quot;...<p>Doesn&#x27;t 7B indicates that it was trained on 7 billion tokens? Or am I misunderstanding the nomenclature?
评论 #35238832 未加载
评论 #35238781 未加载
评论 #35238755 未加载
评论 #35238836 未加载
jakedata大约 2 年前
AI bootstrapping AI is a sci-fi trope that goes back decades. I first encountered it in The Cybernetic Samurai while in high school. While the details differ, the reality is that AI is a catalyst for more of itself.<p>I don&#x27;t remember many books where this ends particularly well. Perhaps the Culture universe could be a survivable outcome. Hopefully we don&#x27;t get Berzerkers first.
评论 #35239559 未加载
评论 #35240670 未加载
xwdv大约 2 年前
Given the high prices of OpenAI offerings it seems it’s better to pirate an AI model before resorting to paying for anything.
评论 #35238699 未加载
评论 #35238687 未加载
评论 #35239442 未加载
评论 #35238875 未加载
amrb大约 2 年前
Anything open will need training and attention to be an openai competitor, tho I&#x27;m happy to see the function of this one: <a href="https:&#x2F;&#x2F;huggingface.co&#x2F;spaces&#x2F;togethercomputer&#x2F;OpenChatKit" rel="nofollow">https:&#x2F;&#x2F;huggingface.co&#x2F;spaces&#x2F;togethercomputer&#x2F;OpenChatKit</a>
alecco大约 2 年前
<a href="https:&#x2F;&#x2F;archive.ph&#x2F;xIKIN" rel="nofollow">https:&#x2F;&#x2F;archive.ph&#x2F;xIKIN</a>
earthboundkid大约 2 年前
I bet you could “exfiltrate” an LLM relatively cheaply by using LLM A to generate training data for LLM B.
评论 #35239533 未加载
UncleOxidant大约 2 年前
Is it accurate to say they were trained for less than $600? Wouldn&#x27;t that just be the finetuning that was done to the already existing LLaMA parameters which likely cost way more than $600 to train?
评论 #35239809 未加载
twblalock大约 2 年前
This is why it&#x27;s not possible to slow down or &quot;stop&quot; AI: once the problems are solved the solutions turn out to be trivial to replicate. All it takes is compute.
评论 #35238797 未加载
评论 #35238807 未加载
EGreg大约 2 年前
I warned about this for years. Finally an article gets it right.<p>Everyone will soon have the equivalent of online nuclear weapons: bot swarms that infiltrate every forum, including this one.
评论 #35239781 未加载