TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

Ask HN: Suggestions for low cost homebrew “HPC” beast?

11 点作者 amuresan将近 7 年前
Over the past few years I&#x27;ve been playing around with the idea of building large number crunching applications for fun and personal research. While the applications I had in mind aren&#x27;t trivially parallel, they should map well to a parallel architecture.<p>There are generally two options for running a big parallel application: 1. going to a commercial provider, which is prohibitively expensive; 2. affiliation to a research lab, which is not the case for me and usually requires a research project for you to get compute time, no playing around with toy application.<p>I was wondering if anyone has experience with building a small homebrew number crunching machine on a budget. The best idea that I have so far is getting a few used server machines with decent GPUs.

6 条评论

anujsharmax将近 7 年前
There is no one thing as &quot;HPC&quot;. It is a lot of normal computer components joined together to form a big &quot;HPC&quot; system. First, identify which kind of scaling you need - are you limited by CPU or memory if you run the code on your PC?<p>On a side note, please don&#x27;t think about buying HPC hardware before actually write the code. You can build the code to solve real problems on your normal PC with a multi core CPU and GPU. Then you can benckmark your parallel code to determine what you acutally need.<p>I could help you out if you can give me more details of what you actually want to achieve.
stuxnet79将近 7 年前
The one factor that distinguishes a typical HPC installation from what we would typically call a &#x27;cloud&#x27; system is the throughpout. HPC installations are designed to provide maximum throughput and this requires very specialized hardware and networking infrastructure (InfiniBand). Cloud installations in contrast make use of widely available commodity hardware (e.g. Raspberry Pis). Unless the tasks you want to do will require an enormous amount of throughput for inter-node messaging within your cluster, then I&#x27;d suggest orienting your plans towards the &#x27;cloud&#x27; model aka using commodity hardware and taking advantage of an open source cloud platform like OpenStack.
closeparen将近 7 年前
If you don’t need to leave it running between computations, this seems like a reasonable EC2&#x2F;GCP use case.<p>University scientific computing labs cost millions of dollars; anything you’re going to build and run at home will be a small-scale replica.
评论 #17640358 未加载
Const-me将近 7 年前
Depends on what exactly you gonna crunch.<p>If you&#x27;re OK running on a GPUs, you essentially need a cryptocurrency mining rig, there lots of articles about them.<p>If you need CPU performance look for used servers on eBay, either complete ones or separate components.
itamarst将近 7 年前
You can get very far with just a normal desktop computer with a bunch of CPUs. E.g. <a href="http:&#x2F;&#x2F;veekaybee.github.io&#x2F;2017&#x2F;03&#x2F;20&#x2F;hadoop-or-laptop&#x2F;" rel="nofollow">http:&#x2F;&#x2F;veekaybee.github.io&#x2F;2017&#x2F;03&#x2F;20&#x2F;hadoop-or-laptop&#x2F;</a>
morphle将近 7 年前
A low cost HPC number cruncher implies you seek the highest performance at its lowest price. The energy used over its lifetime will be the main factor. The capital cost of the processor will probably be less important than the price for the bulk memory.<p>First you should identify the price performance per watt of the systems (processor+memory+network) over their lifetime in your use case. This turns out to be very difficult. The few benchmarks you find online can not be easily compared. The interesting choices are usually just released and have no benchmarks yet. More problematic, benchmarks seldom list the energy and cost of a system<p>every year I try to design the lowest price&#x2F;performance&#x2F;watt system I can find, I spend a few weeks on this. Even with that much effort I have not been able to establish if a cluster of raspberry pi&#x27;s is faster and cheaper than an AMD EPYC with several cpu&#x27;s. Establishing the cost is even harder, if you want to manufacture enough of the systems to benefit from scale. (not a problem in a homebrew system)<p>Currently I guestimate that the best price&#x2F;performance&#x2F;watt system is some mass produced $1-$3 ARM (soon RISC-V) with two DRAM chips and a fast network connection like an FPGA with high speed SERDES link fabric switch. Price per core plus memory must be way below $10. Raspberry pi&#x27;s are no contenders, they only have 200-300 Mbps network for 4 cores to share, that should be several gigabits&#x2F;s to be competitive.<p>A custom build AMD EPYC with several GPU&#x27;s networked together can turn out to be faster and cheaper. The retail price needs to be lowered and you need to build a similiar network fabric switch.<p>You can find better performance if you tailor the processor design to the task. So a system with the best price&#x2F;performance&#x2F;watt will be different for different software.<p>Even cheaper will be systems where you can rebalance part of the compute recources between different programs. The system will can less efficient because the hardware reconfigurability has has a high overhead, but that is offset by being more efficient for one program. You could for example balance transistors between integer, floating point, cache or network-on-chip (NoC).<p>We are currently making prototypes of our own design of a reconfgurable manycore processor with NoC fabric. Around $9 per core in an FPGA. The entry system costs around $80 and $500.<p>We plan to build this as an ASIC, the price will then drop to $1 per core including DRAM.<p>Even cheaper will be if we not slice the wafer into 22000 chips but leave it whole. You get over 100.000 cores with little memory with petabits&#x2F;s network for less than $6000. Half the wafer is reconfigurable logic that can be reprogrammed at runtime as GPU, TPU, CPU or any other custom optimisation. The energy cost of the wafer can be zero if you use the wafer as a water heating element and lower if you only run it on solar PV during the day. (If you share two wafers with a user on the night time side of the earth, you both can have 24 hour computing on $0,02 per kWh of solar PV). We can make a 180nm $500 version and a $1 version but they would not be as good as the 28nm $6000 version. A future 7nm wafer scale version might never be cheaper than the 28nm version, we will have to wait and see.<p>An silicon optical network on the wafer (also $6000) would allow two or more wafers to be networked at several terabits&#x2F;s. This overcomes its small memory problem.<p>Think of it, you want to get rid of any overhead in the system, like pcb, chip package, connectors, cables. You want to put everything on a single large chip, the whole wafer. Because we have different CMOS technology for DRAM or processors, we wind up with needing one wafer.<p>We are confident in 30 years we can grow the wafer(or 3D block) and the solar panel from CO2.
评论 #17644949 未加载
评论 #17638141 未加载