TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

Ask HN: Is distributed LLM training in browsers (WebRTC and WebGPU) possible?

4 点作者 trekhleb6 个月前

5 条评论

trekhleb6 个月前
What I mean is training something like GPT-3 in a distributed manner using a large number of regular browsers or laptops with average WebGPU support&#x2F;power and WebRTC for communication.<p>Does it even make sense to ask this? Is it reasonable or feasible?<p>I understand there are many nuances, such as the size and source of the training data, the size of the model (which would be too large for any browser to handle), network overhead, and the challenge of merging all the pieces together, among others. However, speculative calculations suggest that GPT-3 required around 3x10^22 FLOPs, which might (very speculatively) be equivalent to about 3,000 regular GPUs, each with an average performance of 6 TFLOPs, training it for ~30 days (which also sounds silly, I understand).<p>Of course, these are naive and highly speculative calculations that don’t account for whether it’s even possible to split the dataset, model, and training process into manageable pieces across such a setup.<p>But if this direction is not totally nonsensical, does it mean that even with a tremendous network overhead there is a huge potential for scaling (there are potentially a lot of laptops connected to the internet that potentially and voluntary could be used for training)?
soheil6 个月前
This is the promise of WebGPU. Once all major browsers fully adopt it it&#x27;ll be a game change for any ML training. Apple Silicon is the giant no one is talking about training is almost exclusively done on NVDA chips. Cost per TFlop is the lowest on Apple chips both operational and acquisition (mac mini).
astlouis446 个月前
Client-side inference will be a bigger opportunity for AI in the browser than client-side training, in my opion.
gtsop6 个月前
Possible? Yes. Practical? No<p>It is possible because training is already being done using distributed methods in the datacenter. It is not practical because of the sheer volume of network delay piled on top of the computation
xyzsparetimexyz6 个月前
Distributed training in general isn&#x27;t practical.