TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

Try SambaNova chat: 1T param LLM, 500 tokens/SEC

1 点作者 germanjoey大约 1 年前

1 comment

germanjoey大约 1 年前
We&#x27;re showing off our 1.05T param Composition of Experts LLM! It&#x27;s 150 experts running on 1 node consisting of 8 SN40L RDU chips.<p>Each of our nodes has a huge amount of DDR attached, in addition to copious amounts of on-chip HBM and SRAM. This allows the system to switch between a variety of different models of different sizes and architectures at lightning speed. A highlight is one based on Llama2 7b, similar to the Groq demo, but executing with bf16&#x2F;fp32 instead of int8. (And using only 8 chips instead of 568!)