TE
TechEcho
Home24h TopNewestBestAskShowJobs
GitHubTwitter
Home

TechEcho

A tech news platform built with Next.js, providing global tech news and discussions.

GitHubTwitter

Home

HomeNewestBestAskShowJobs

Resources

HackerNews APIOriginal HackerNewsNext.js

© 2025 TechEcho. All rights reserved.

My Experience and Advice for Using GPUs in Deep Learning: Which GPU to get

158 pointsby michaelnalmost 7 years ago

9 comments

lern_too_spelalmost 7 years ago
The &quot;I have almost no money&quot; recommendation should include Colab. <a href="https:&#x2F;&#x2F;medium.com&#x2F;deep-learning-turkey&#x2F;google-colab-free-gpu-tutorial-e113627b9f5d" rel="nofollow">https:&#x2F;&#x2F;medium.com&#x2F;deep-learning-turkey&#x2F;google-colab-free-gp...</a><p>Somebody who has almost no money isn&#x27;t going to be able to equip a desktop with a GTX 1050 Ti ($175), fast disk ($50), and RAM ($50) on an entry level cpu&#x2F;motherboard&#x2F;power supply&#x2F;case&#x2F;monitor&#x2F;peripherals ($300) and pay for the electricity used during training. Colab can be accessed from a free public computer or a cheap Chromebook ($200).
评论 #17816984 未加载
评论 #17815650 未加载
评论 #17815364 未加载
评论 #17817383 未加载
sabalabaalmost 7 years ago
The 2080Ti numbers are likely going to be a lot lower than that.<p>We’ve benched the 1080Ti vs the Titan V and the Titan V is nowhere near 2x faster at training than the 1080Ti as suggested in that graph. We observed a 30% to 40% speedup during our benchmarking:<p><a href="https:&#x2F;&#x2F;deeptalk.lambdalabs.com&#x2F;t&#x2F;benchmarking-the-titan-v-volta-gpu-with-tensorflow&#x2F;108" rel="nofollow">https:&#x2F;&#x2F;deeptalk.lambdalabs.com&#x2F;t&#x2F;benchmarking-the-titan-v-v...</a><p>This is consistent with the 32% increase in FP32 flops from 11.3TFlops for the 1080Ti to 15TFlops for the Titan V. Additional speedups can be explained by the increase in memory bandwidth for HBM2 and the mixed precision fused multiply adds provided by the TensorCores.<p>Thus, given the quoted 13Tflop numbers for the 2080Ti, I would expect the 2080Ti to present something more like a 15-20% speedup over the 1080Ti. So 2080Ti is less bang for your buck. But benchmarking is the only way to tell what’s better on a FLOPS&#x2F;$ basis.
评论 #17817399 未加载
ageitgeyalmost 7 years ago
This is a great article and I highly respect his opinions.<p>However, since you are probably eagerly reading this to see how fast the new RTX cards are, so you should know upfront that the numbers he has so far are just estimates based on specs:<p>&gt; Note that the numbers for the RTX 2080 and RTX 2080 Ti should be taken with a grain of salt since no hard performance numbers existed. I estimated performance according to a roofline model of matrix multiplication and convolution under this hardware together with Tensor Core benchmarks from the V100 and Titan V.
评论 #17813704 未加载
评论 #17812294 未加载
pirocksalmost 7 years ago
Seems down for me:<p><a href="https:&#x2F;&#x2F;web.archive.org&#x2F;web&#x2F;20180821173206&#x2F;http:&#x2F;&#x2F;timdettmers.com&#x2F;2018&#x2F;08&#x2F;21&#x2F;which-gpu-for-deep-learning&#x2F;" rel="nofollow">https:&#x2F;&#x2F;web.archive.org&#x2F;web&#x2F;20180821173206&#x2F;http:&#x2F;&#x2F;timdettmer...</a>
scottlegrand2almost 7 years ago
The biggest advance here is that Nvidia has produced a consumer card that has all the high-end deep-learning features. This was missing in both the Pascal and Volta Generations even though in Pascal fp32 was full power. I think the TPU scared them and that&#x27;s a good thing.
syntaxingalmost 7 years ago
Hacker news hug of death? Anyone here have any experience using AMD cards with something like PlaidML? I have a 1050Ti SSC but I&#x27;m starting to feel the limitation as my complexity grows. But getting a 1080 is a bit out of my budget right now. I&#x27;m tempted to get the new Vega 56 released recently.
评论 #17813842 未加载
fermienricoalmost 7 years ago
The cost&#x2F;performance plot - shouldn&#x27;t it be &quot;Lower is better&quot;? It says &quot;Higher is better&quot;.<p>Lower value would indicate lower cost per unit level of performance.<p>It should be &quot;Lower is better&quot; or the plot needs to say &quot;Performance&#x2F;Cost&quot;. Am I missing something?
评论 #17817368 未加载
评论 #17814970 未加载
评论 #17813655 未加载
dostresalmost 7 years ago
An open question for me is the performance of two 2080tis using NVLink as one virtual GPU. I imagine it’ll be close to linear, but I’ll be interested to know for sure.
评论 #17816420 未加载
KayLalmost 7 years ago
Good article, but as a new learner, I&#x27;m interested in (your experiences on) how much time taken for the common task to train a model? 1min vs 2mins, probably I will get a cheaper GPU but if there&#x27;s 5h vs 10h or 1 day vs 2 days, I&#x27;d save more money for one with good performance