TE
TechEcho
Home24h TopNewestBestAskShowJobs
GitHubTwitter
Home

TechEcho

A tech news platform built with Next.js, providing global tech news and discussions.

GitHubTwitter

Home

HomeNewestBestAskShowJobs

Resources

HackerNews APIOriginal HackerNewsNext.js

© 2025 TechEcho. All rights reserved.

GPT-4: 8 x 220B experts trained with different data/task distributions

47 pointsby MasterScratalmost 2 years ago

4 comments

SheinhardtWigCoalmost 2 years ago
As a heavy user of GPT-4 (I&#x27;m working on a plugin), reading this felt like a puzzle piece being dropped into place.<p>Maybe this is just confirmation bias, but yeah, trying to push the model&#x27;s capabilities is like working with a committee of brilliant minds chaired by an idiot.<p>Also, I can see why they kept this secret. Competitors just shaved months off their R&amp;D timelines.
euclaisealmost 2 years ago
The only paper that I could find using an approach with fully separated experts like this is <a href="https:&#x2F;&#x2F;arxiv.org&#x2F;pdf&#x2F;2208.03306.pdf" rel="nofollow noreferrer">https:&#x2F;&#x2F;arxiv.org&#x2F;pdf&#x2F;2208.03306.pdf</a>
swyxalmost 2 years ago
the source podcast that this came from: <a href="https:&#x2F;&#x2F;news.ycombinator.com&#x2F;item?id=36407269">https:&#x2F;&#x2F;news.ycombinator.com&#x2F;item?id=36407269</a>
adeonalmost 2 years ago
Is this an actually confirmed detail or just something George Hotz speculated? How credible is it?
评论 #36419526 未加载