TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

Building a deep learning rig

164 点作者 dvcoolarun大约 1 年前

16 条评论

Yenrabbit大约 1 年前
Note that they shared part two recently: <a href="https:&#x2F;&#x2F;samsja.github.io&#x2F;blogs&#x2F;rig&#x2F;part_2&#x2F;" rel="nofollow">https:&#x2F;&#x2F;samsja.github.io&#x2F;blogs&#x2F;rig&#x2F;part_2&#x2F;</a><p>For those talking about breakeven points and cheap cloud compute, you need to factor in the mental difference it makes running a test locally (which feels free) vs setting up a server and knowing you&#x27;re paying per hour it&#x27;s running. Even if the cost is low, I do different kinds of experiments knowing I&#x27;m not &#x27;wasting money&#x27; every minute the GPU sits idle. Once something is working, then sure scaling up on cheap cloud compute makes sense. But it&#x27;s really, really nice having local compute to get to that state.
评论 #39493925 未加载
评论 #39498648 未加载
评论 #39496124 未加载
评论 #39494296 未加载
abra0大约 1 年前
I was thinking of doing something similar, but I am a bit sceptical about how the economics on this works out. On vast.ai renting a 3x3090 rig is $0.6&#x2F;hour. The electricity price of operating this in e.g. Germany is somewhere about $0.05&#x2F;hour. If the OP paid 1700 EUR for the cards, the breakeven point would be around (haha) 3090 hours in, or ~128 days, assuming non-stop usage. It&#x27;s probably cool to do that if you have a specific goal in mind, but to tinker around with LLMs and for unfocused exploration I&#x27;d advise folks to just rent.
评论 #39492112 未加载
评论 #39492247 未加载
评论 #39492256 未加载
评论 #39492578 未加载
评论 #39493813 未加载
评论 #39492538 未加载
评论 #39493574 未加载
评论 #39498226 未加载
评论 #39498671 未加载
评论 #39491845 未加载
评论 #39492991 未加载
评论 #39493710 未加载
评论 #39495820 未加载
infogulch大约 1 年前
I&#x27;m eyeing Tinybox as a deep learning rig.<p><a href="https:&#x2F;&#x2F;tinygrad.org&#x2F;" rel="nofollow">https:&#x2F;&#x2F;tinygrad.org&#x2F;</a><p><a href="https:&#x2F;&#x2F;twitter.com&#x2F;__tinygrad__&#x2F;status&#x2F;1760988080754856210" rel="nofollow">https:&#x2F;&#x2F;twitter.com&#x2F;__tinygrad__&#x2F;status&#x2F;1760988080754856210</a>
评论 #39491777 未加载
评论 #39491751 未加载
评论 #39498269 未加载
评论 #39491648 未加载
bick_nyers大约 1 年前
Somewhat tangential question, but I&#x27;m wondering if anyone knows of a solution (or Google search terms for this):<p>I have a 3U supermicro server chassis that I put an AM4 motherboard into, but I&#x27;m looking at upgrading the Mobo so that I can run ~6 3090s in it. I don&#x27;t have enough physical PCIE slots&#x2F;brackets in the chassis (7 expansion slots), so I either need to try to do some complicated liquid cooling setup to make the cards single slot (I don&#x27;t want to do this), or I need to get a bunch of riser cables and mount the GPU above the chassis. Is there like a JBOD equivalent enclosure for PCIE cards? I don&#x27;t really think I can run the risers out the back of the case, so I&#x27;ll likely need to take off&#x2F;modify the top panel somehow. What I&#x27;m picturing in my head is basically a 3U to 6U case conversion, but I&#x27;m trying to minimize cost (let&#x27;s say $200 for the chassis&#x2F;mount component) as well as not have to cut metal.
评论 #39495016 未加载
评论 #39495008 未加载
kaycebasques大约 1 年前
I really enjoy and am inspired by the idea that people like Dettmer (and probably this Samsja person) are the spiritual successors to homebrew hackers in the 70s and 80s. They have pretty intimate knowledge of many parts of the whole goddamn stack, from what&#x27;s going on in each hardware component, to how to assemble all the components into a rig, up to all the software stuff: algorithms, data, orchestration, etc.<p>Am also inspired by embedded developers for the same reason
neilv大约 1 年前
For large VRAM models, what about selling one of the 3090s, and putting the money towards an NVLink and a motherboard with two x16 PCIe slots (and preferably spaced so you don&#x27;t need riser cables)?
评论 #39494457 未加载
评论 #39494036 未加载
评论 #39493840 未加载
评论 #39493784 未加载
Uehreka大约 1 年前
&gt; I just got my hands on a mining rig with 3 rtx 3090 founder edition for the modest sum of 1.7k euros.<p>I would prefer a tutorial on how to do this.
whoisthemachine大约 1 年前
I&#x27;ve been slowly expanding my HTPC&#x2F;media server into a gaming server and box for running LLMs (and possibly diffusion models?) locally for playing around with. I think it&#x27;s becoming clear that the future of LLM&#x27;s will be local!<p>My box has a Gigabyte B450M, Ryzen 2700X, 32GB RAM, Radeon 6700XT (for gaming&#x2F;streaming to steam link on Linux), and an &quot;old&quot; Geforce GTX 1650 with a paltry 6GB of RAM for running models on. Currently it works nicely with smaller models on ollama :) and it&#x27;s been fun to get it set up. Obviously, now that the software is running I could easily swap in a more modern NVidia card with little hassle!<p>I&#x27;ve also been eyeing the b450 steel legend as a more capable board for expansion than the Gigabyte board, this article gives me some confidence that it is a solid board.
smokeydoe大约 1 年前
Does anyone have any good recommendations for an epyc server grade motherboard that can use 3x3090? My current motherboard (strix trx40-xe) has memory issues now. 2 slots cause boot errors no matter what memory is inserted. I plan to sell the threadripper. Other option is to just swap out the current motherboard with a trx zenith extreme but I feel server grade would be better at this point after experiencing issues. Is supermicro worth it?
评论 #39494109 未加载
评论 #39493606 未加载
评论 #39496431 未加载
评论 #39494303 未加载
0x20cowboy大约 1 年前
If you would like to put Kubernetes on top of this kind of setup this repo is helpful <a href="https:&#x2F;&#x2F;github.com&#x2F;robrohan&#x2F;skoupidia">https:&#x2F;&#x2F;github.com&#x2F;robrohan&#x2F;skoupidia</a><p>The main benefit is you can shut off nodes entirely when not using them, and then when you turn them back on they just rejoin the cluster.<p>It also helps managing different types of devices and workloads (tpu vs gpu vs cpu)
评论 #39495806 未加载
akasakahakada大约 1 年前
Just sharing.<p>2 x RTX4090 workstation guide<p>You can put two aircooled 4090 in the same ATX case if you do enough research.<p><a href="https:&#x2F;&#x2F;github.com&#x2F;eul94458&#x2F;Memo&#x2F;blob&#x2F;main&#x2F;dual_rtx4090workstation_for_machine_learning_202401.md">https:&#x2F;&#x2F;github.com&#x2F;eul94458&#x2F;Memo&#x2F;blob&#x2F;main&#x2F;dual_rtx4090works...</a>
cyanydeez大约 1 年前
just ordered a 15k thread ripper platform because it&#x27;s the only way to cheaply maximize the pcie16x bottleneck. the mining rigs are neat because the space you need for consumer GPU is a big issue.<p>those rigs need pcie riser slots that are also limited.<p>looks like the primary value is the rig and the cards. they&#x27;ll need another 1-2k for a thread ripper and then the riser slots.
评论 #39492099 未加载
评论 #39493936 未加载
nirav72大约 1 年前
This is nice. I would’ve used one of those ETH mining cases that support multiple GPUs. Ebay has them $100-150 these days.
whimsicalism大约 1 年前
I strongly, strongly suspect most people doing this are significantly short of the breakeven prices for transitioning from cloud 3090s.<p>inb4 there are no cloud 3090s: yes there are, just not in formal datacenters
评论 #39493736 未加载
jeffybefffy519大约 1 年前
Are m1&#x2F;m2&#x2F;m3 max mac&#x27;s any good for this?
评论 #39495655 未加载
gigatexal大约 1 年前
I thought this looked like a cryptocurrency miner. Seems the crypto to AI pivot is legit happening. And good. Would rather we boiled the oceans for something marginally more valuable than in-game tokens we traded for fiat funds in this video game we call life.