TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

Intellect-2 Release: The First 32B Model Trained Through Globally Distributed RL

201 点作者 Philpax10 天前

16 条评论

throwanem10 天前
<i>There&#x27;s</i> a name and a logo. &quot;Hubris&quot; feels slightly beggared. <a href="https:&#x2F;&#x2F;en.m.wikipedia.org&#x2F;wiki&#x2F;The_Metamorphosis_of_Prime_Intellect" rel="nofollow">https:&#x2F;&#x2F;en.m.wikipedia.org&#x2F;wiki&#x2F;The_Metamorphosis_of_Prime_I...</a>
评论 #43961570 未加载
评论 #43963004 未加载
refulgentis10 天前
I guess I&#x27;m bearish?<p>It&#x27;s not that they <i>trained a new model</i>, but <i>they took an existing model</i> and RL&#x27;d it a bit?<p>The scores are very close to QwQ-32B, and at the end:<p>&quot;Overall, as QwQ-32B was already extensively trained with RL, it was difficult to obtain huge amounts of generalized improvement on benchmarks beyond our improvements on the training dataset. To see stronger improvements, it is likely that better base models such as the now available Qwen3, or higher quality datasets and RL environments are needed.&quot;
评论 #43959398 未加载
评论 #43959451 未加载
评论 #43961411 未加载
iTokio10 天前
It’s interesting that it does something useful (training a LLM) without trust and in a decentralized way.<p>Maybe this could be used as proof of work? To stop wasting computing resources in crypto currencies and get something useful as a byproduct.
评论 #43960631 未加载
评论 #43960685 未加载
评论 #43963132 未加载
评论 #43960171 未加载
评论 #43959974 未加载
评论 #43960031 未加载
评论 #43960375 未加载
评论 #43961433 未加载
3abiton10 天前
This is rather exciting! I see the future of Co-op models made by a community of experts on a specific field that would still allow them to be competitive with &quot;AI monopolies&quot;. Maybe not all hope is lost!
Thomashuet10 天前
Summary: We&#x27;ve use the most complexest, buzzwordiest training infrastructure to increase the performance of our base model by a whopping 0.5% (±1%).
评论 #43961426 未加载
评论 #43964118 未加载
danielhanchen10 天前
I made some GGUFs at <a href="https:&#x2F;&#x2F;huggingface.co&#x2F;unsloth&#x2F;INTELLECT-2-GGUF" rel="nofollow">https:&#x2F;&#x2F;huggingface.co&#x2F;unsloth&#x2F;INTELLECT-2-GGUF</a><p>.&#x2F;llama.cpp&#x2F;llama-cli -hf unsloth&#x2F;INTELLECT-2-GGUF:Q4_K_XL -ngl 99<p>Also it&#x27;s best to read <a href="https:&#x2F;&#x2F;docs.unsloth.ai&#x2F;basics&#x2F;tutorial-how-to-run-qwq-32b-effectively">https:&#x2F;&#x2F;docs.unsloth.ai&#x2F;basics&#x2F;tutorial-how-to-run-qwq-32b-e...</a> on sampling issues for QwQ based models.<p>Or TLDR, use the below settings:<p>.&#x2F;llama.cpp&#x2F;llama-cli -hf unsloth&#x2F;INTELLECT-2-GGUF:Q4_K_XL -ngl 99 --temp 0.6 --repeat-penalty 1.1 --dry-multiplier 0.5 --min-p 0.00 --top-k 40 --top-p 0.95 --samplers &quot;top_k;top_p;min_p;temperature;dry;typ_p;xtc&quot;
abtinf10 天前
Does this have anything to do with The Metamorphosis Of Prime Intellect, or did they just abuse the name and the cover art?
评论 #43959483 未加载
esafak10 天前
How are they ensuring robustness against adversarial responses?
评论 #43959233 未加载
schneehertz10 天前
I used to have an idea related to science fiction novels that artificial intelligence could aggregate computing power through the network to perform ultra-large-scale calculations, thereby achieving strong artificial intelligence. Reality will also develop in this way, which is very interesting
mountainriver10 天前
Awesome work this team is doing. Globally distributed MoE could have real legs
quantumwoke10 天前
Wonder what the privacy story is like. Enterprises don&#x27;t usually like broadcasting their private data across a freely accessible network.
评论 #43959511 未加载
bwfan12310 天前
The most interesting thing I see is the productization of the diloco work done here [1]. If someone can make this scale, then we can say goodbye to expensive backend networking and mainframe-like AI training machinery.<p>[1] <a href="https:&#x2F;&#x2F;arxiv.org&#x2F;abs&#x2F;2311.08105" rel="nofollow">https:&#x2F;&#x2F;arxiv.org&#x2F;abs&#x2F;2311.08105</a>
ikeashark10 天前
I wonder why they randomly noted a torch-compile vs non torch-compile figure where torch-compile degraded model performance. What made it degrade? It seems to only appear in one figure and nowhere else.
评论 #43968961 未加载
ndgold10 天前
Pretty badass
Mougatine10 天前
very cool work!
jumploops10 天前
Congrats to the team on the launch!<p>Personal story time: I met a couple of their engineers at an event a few months back. They mentioned they were building a distributed training system for LLMs.<p>I asked them how they were building it and they mentioned Python. I said something along the lines of “not to be the typical internet commenter guy, but why aren’t you using something like Rust for the distributed system parts?”<p>They mumbled something about Python as the base for all current LLMs, and then kinda just walked away…<p>From their article: &gt; “Rust-based orchestrator and discovery service coordinate permissionless workers”<p>Glad to see that I wasn’t entirely off-base :)
评论 #43964163 未加载
评论 #43960455 未加载
评论 #43959547 未加载