I was just saying to a colleague the day before this announcement that the inevitable consequence of the popularity of <i>large</i> language models will be GPUs with more memory.<p>Previously, GPUs were designed for gamers, and no game really "needs" more than 16 GB of VRAM. I've seen reviews of the A100 and H100 cards saying that the 80GB is ample for even the most demanding usage.<p>Now? Suddenly GPUs with 1 TB of memory could be <i>immediately</i> used, at scale, by deep-pocket customers happy to throw their entire wallets at NVIDIA.<p>This new H100 NVL model is a Frankenstein's monster stitched together from whatever they had lying around. It's a desperate move to corner the market early as possible. It's just the beginning, a preview of the times to come.<p>There will be a new digital moat, a new capitalist's empire, built upon on the scarcity of cards "big enough" to run models that nobody but a handful of megacorps can afford to train.<p>In fact, it won't be enough to restrict access by making the models expensive to train. The real moat will be models too expensive to run. Users will have to sign up, get API keys, and stand in line.<p>"Safe use of AI" my ass. Safe profits, more like. Safe monopolies, safe from competition.