TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

RTX3090 TensorFlow, NAMD and HPCG Performance on Linux

70 点作者 polymorph1sm超过 4 年前

9 条评论

SloopJon超过 4 年前
Enroot sounds interesting:<p><a href="https:&#x2F;&#x2F;github.com&#x2F;NVIDIA&#x2F;enroot" rel="nofollow">https:&#x2F;&#x2F;github.com&#x2F;NVIDIA&#x2F;enroot</a><p>It basically returns containers to their chroot origins, promising &quot;no performance overhead.&quot; I&#x27;m looking forward to more posts on that.
评论 #24664711 未加载
评论 #24665005 未加载
kernelsanderz超过 4 年前
I was originally appalled at the software limiting. But according to Tim Dettmers who has a solid record of predicting and comparing NVIDIA cards for deep learning performance, it&#x27;s not really a big deal.<p>You can read his analysis here: <a href="https:&#x2F;&#x2F;timdettmers.com&#x2F;2020&#x2F;09&#x2F;07&#x2F;which-gpu-for-deep-learning&#x2F;" rel="nofollow">https:&#x2F;&#x2F;timdettmers.com&#x2F;2020&#x2F;09&#x2F;07&#x2F;which-gpu-for-deep-learni...</a><p>and his tweet about this here: <a href="https:&#x2F;&#x2F;twitter.com&#x2F;Tim_Dettmers&#x2F;status&#x2F;1311354118514982912" rel="nofollow">https:&#x2F;&#x2F;twitter.com&#x2F;Tim_Dettmers&#x2F;status&#x2F;1311354118514982912</a><p>Essentially from my understanding it&#x27;s memory bandwidth which is the real critical path on performance in most cases. The previous generation of Turing cards had more compute than was necessary so they were an underutilized resource.<p>Also, this Puget benchmark is using an older version of the CUDA drivers. I believe performance is much better in CUDA 11.1.<p>This new benchmark which is running on the latest CUDA seems to confirm Tim&#x27;s numbers: <a href="https:&#x2F;&#x2F;www.evolution.ai&#x2F;post&#x2F;benchmarking-deep-learning-workloads-with-tensorflow-on-the-nvidia-geforce-rtx-3090" rel="nofollow">https:&#x2F;&#x2F;www.evolution.ai&#x2F;post&#x2F;benchmarking-deep-learning-wor...</a>
bitL超过 4 年前
When I first read the review, I couldn&#x27;t understand why the author was mentioning that FP16 will surely be improved by new drivers despite not understanding that FP16 TFlops are exactly the same as FP32, tensor cores were nerfed and FP32 accumulate set to 0.5x speed, instead of Titan RTX&#x27;s 1x. I&#x27;d say the results are as good as it gets, if you want a better performance, wait for Ampere Titan or Quadro.
评论 #24662027 未加载
评论 #24664125 未加载
ilzmastr超过 4 年前
I&#x27;m looking forward to multi-gpu tests!<p>Would be good to see if it is worth upgrading x4 and x8 setups.<p>Single gpu upgrade being worth it is a no-brainer. Launch price of the 30xx cards is lower that then purchase-able price of the two comparison cards!<p>If only you could buy them though. The only microcenter in PA got 15 of each 30xx on respective launch days.<p>If anyone knows how many of these cards are being produced please do share.
komuher超过 4 年前
Its without XLA and without cudnn and cuda 11.1 for ga102 lets wait for proper drivers to see full results :P
optimalsolver超过 4 年前
Is there any reason NVIDIA aren&#x27;t selling A100s as individual cards?
评论 #24663727 未加载
rkwasny超过 4 年前
They just have the wrong ptxas - needs Cuda 11.1 and properly set $PATH.
ngcc_hk超过 4 年前
Guess that is maximum many of us can afford. Hence features that are missing from A100 is a bit moot. But the update we are wait. Still based on what we saw 3090 really does not worth the money. Still 24gb is 24gb.
评论 #24662778 未加载
评论 #24659978 未加载
unstatusthequo超过 4 年前
Neat, would love to try it but can’t buy the card anywhere. Another shitty Nvidia launch.