Big Numbers are good for headline. But it doesn't put anything in context.<p>Die Size is 645mm^2 on a 7nm. This is important because we know the reticle limit which is around ~800mm^2.<p>The Nvidia AI Chip has 54 billion transistors with a die size of 826 mm2 on 7nm.<p>I recently saw a Ted Talk, If Content is King, then Context is God. I think it capture everything that is wrong in today's society.
The article starts with this statement:<p><pre><code> Artificial intelligence (AI) has
seen a broad adoption over the past
couple of years.
</code></pre>
And continues:<p><pre><code> At Tesla, who as many know is a
company that works on electric
and autonomous vehicles, AI has
a massive value to every aspect
of the company's work.
</code></pre>
Who is writing like this? And why?<p>What would Tom's Hardware lose if they left out this type of cheap fillwords?<p>Should I also start writing like this?<p>Is this type of "reader hostile writing" a new thing or have newspapers always written like this?<p>These are not rhetorical questions. I am honestly confused.
This is made using TSMC's 7nm fab process so surly the number of transistors in this chip is either enabled or limited by that process, isn't it?<p>Honest question, how much is chip design a factor separate to fab process?
I'm always curious about the decision-making progress when someone decides to make their own ASIC when there are somewhat reasonable commercial alternatives. What was the advantage here for Tesla?
Tesla actually has a lot of expertise in chip design in Pete Bannon and formerly Jim Keller. I think most people know who Jim Keller is, but if not you can read his wikipedia[1]. Pete Bannon is also an industry giant and worked with Jim Keller at PA Semi and subsequently Apple on their A series chips. These two have decades of experience designing chips that went into tens of millions of devices. Tesla’s FSD computer is in hundreds of thousands of cars. They know what they’re doing.<p><a href="https://en.wikipedia.org/wiki/Jim_Keller_(engineer)" rel="nofollow">https://en.wikipedia.org/wiki/Jim_Keller_(engineer)</a>
Question: how many chips does Tesla need to buy in order to get a reasonable unit price per chip? Obviously <10k is too small, but is 100k reasonable? 1M?
The whole point of the die on silicone seems to be that this maximizes the interface bandwidth and minimize latency between the dies. If this true the next step would be to bring the multi die modules as close as possible in three dimensions to ultimately build a borg-cube like structure in zero-g with a power source at its core.
I wonder how their neural network structures informed the hardware design, such as the dimensions of tensor products. Or is Dojo trying for as general purpose ML as possible? I imagine there is a tension between software and hardware teams where Karpathy's team is always changing things while the hardware team wants specs/reqs.<p>The "tiles of tiles" chip architecture seems like an Elon-obvious, let's just scale what we have approach. Do their neural networks map to that multiscale tiling well?
Non-hardware person here. How does the D1 compare to Cerebras WSE-2 wafer chip with 2.6 trillion transistors?<p>The WSE2 is much larger obviously, but I would also think it can result in a large performance boost given everything is on a single chip.