TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

A Really Big Computer

60 点作者 cl3m将近 2 年前

12 条评论

bastawhiz将近 2 年前
Maybe I'm overly critical, but if fixing the Twitter search box is too hard, maybe buying solar (and batteries), fabbing chips, and building a giant data center to replicate the current generation of what another company already built is foolhardy.
评论 #37064184 未加载
评论 #37073203 未加载
评论 #37071511 未加载
gary_0将近 2 年前
He doesn&#x27;t mention yield, which I hear is currently pretty bad on 3nm. And &quot;just route around the bad silicon&quot; might be easier said than done; widespread defects might create too many roadblocks to efficiently move data around your wafer-chip.<p>If that&#x27;s not factored into his 3450 wafer estimate, it could be double that.
评论 #37065619 未加载
20wenty将近 2 年前
Say what you want about him, but George is one of the few people I know that can actually go a mile wide and a mile deep into any field they choose. And he does most of his digging live on stream, so he makes an easy punching bag. Do I think he&#x27;ll successfully raise a $400M round on a $2B valuation? No, something else will distract him before that - but I&#x27;m going to enjoy watching him (and pulling for him) regardless.
评论 #37070464 未加载
w10-1将近 2 年前
Brainstorming can help identify limiting assumptions.<p>Mostly compute has piggy-backed off consumer-scale production (e.g., GPU&#x27;s repurposed for crypto).<p>The suggestion is that an AI model can justify few-shot chip production.<p>His proposal is for development, i.e., to build the model, and depends mostly on such models being qualitatively better.<p>It seems more likely that chips would be built to offer model processing, instead of forcing users into a service (with its risk of confidentiality and IP leaks). To get GPT-100, you&#x27;d incorporate the chip into your device -- and then know for sure that nothing could leak. That eliminates the primary transaction cost for AI compute: the risk.<p>Which presents the question: does anyone know of research or companies working on such chip models?
leblancfg将近 2 年前
I will refer to this as the moment when LLM hype jumped the shark.
评论 #37065594 未加载
nickdothutton将近 2 年前
How would you distribute clock signal around that wafer-scale GPU? OR is he simply suggesting you buy the whole wafer and litho standard GPUs out of it? Apologies if this is a stupid question.
petesoper将近 2 年前
How does he cool those wafers?
评论 #37066895 未加载
nullc将近 2 年前
How do these goals set out the needed ratios of memory to compute to interconnect bandwidth?<p>An ideal machine designed to train GPT4 in a day is likely very different to the ideal machine to train 50 GPT4s as once over a few weeks, which is very different from the ideal machine to train a model 100x bigger than GPT4 (perhaps the most interesting).
Havoc将近 2 年前
Not sure about that computer plan but I do enjoy his brand of entertainment.<p>Also really hoping he makes more progress on amd ml
nullc将近 2 年前
Where is the budget line item for defending it against bombing by crazed anti-ai doomsday cultists?
jebarker将近 2 年前
Wyoming seems like a good place to build such a thing.
评论 #37074100 未加载
dealuromanet将近 2 年前
This is really cool. Godspeed!