TE
科技回声
首页
24小时热榜
最新
最佳
问答
展示
工作
中文
GitHub
Twitter
首页
Llama 405B 506 tokens/second on an H200
21 点
作者
moondistance
7 个月前
3 条评论
EgoIncarnate
7 个月前
Collapse
not "an H200", "In the table above, tensor parallelism is compared to pipeline parallelism with each across eight GPUs"
评论 #41833897 未加载
7e
7 个月前
Collapse
And this is why nobody submits MLPerf against NVIDIA.
评论 #41833932 未加载
moondistance
7 个月前
Significant further optimizations. FP8!