According to the company, the new chip will enable training of AI models with up to 24 trillion parameters. Let me repeat that, in case you're as excited as I am: <i>24. Trillion. Parameters.</i> For comparison, the largest AI models currently in use have around 0.5 trillion parameters, around 48x times smaller.<p>Each parameter is a <i>connection between artificial neurons</i>. For example, inside an AI model, a linear layer that transforms an input vector with 1024 elements to an output vector with 2048 elements has 1024×2048 = ~2M parameters in a weight matrix. Each parameter specifies by how much each element in the input vector contributes to or subtracts from each element in the output vector. Each output vector element is a weighted sum (AKA a linear combination), of each input vector element.<p>A human brain has an estimated 100-500 trillion synapses connecting biological neurons. Each synapse is quite a complicated biological structure[a], but if we oversimplify things and assume that every synapse can be modeled as a single parameter in a weight matrix, then the largest AI models in use today have approximately 100T to 500T ÷ 0.5T = 200x to 1000x fewer connections between neurons that the human brain. If the company's claims prove true, this new chip will enable training of AI models that have only 4x to 20x fewer connections that the human brain.<p>We sure live in interesting times!<p>---<p>[a] <a href="https://en.wikipedia.org/wiki/Synapse" rel="nofollow">https://en.wikipedia.org/wiki/Synapse</a>