AI memory is sold out, causing an unprecedented surge in prices
Eugene Mymrin | Moment | Getty Images
All computing devices require a part called memory, or RAM, for short-term data storage, but this year, there won’t be enough of these essential components to meet worldwide demand.
That’s because companies like Nvidia, Advanced Micro Devices and Google need so much RAM for their artificial intelligence chips, and those companies are the first ones in line for the components.
Three primary memory vendors — Micron, SK Hynix and Samsung Electronics — make up nearly the entire RAM market, and their businesses are benefitting from the surge in demand.
“We have seen a very sharp, significant surge in demand for memory, and it has far outpaced our ability to supply that memory and, in our estimation, the supply capability of the whole memory industry,” Micron business chief Sumit Sadana told CNBC this week at the CES trade show in Las Vegas.
Micron’s stock is up 247% over the past year year, and the company reported that net income nearly tripled in the most recent quarter. Samsung this week said that it expects its December quarter operating profit to nearly triple as well. Meanwhile, SK Hynix is considering a U.S. listing as its stock price in South Korea surges, and in October, the company said it had secured demand for its entire 2026 RAM production capacity.
Now, prices for memory are rising.
TrendForce, a Taipei-based researcher that closely covers the memory market, this week said it expects average DRAM memory prices to rise between 50% and 55% this quarter versus the fourth quarter of 2025. TrendForce analyst Tom Hsu told CNBC that type of increase for memory prices was “unprecedented.”
Three-to-one basis
Chipmakers like Nvidia surround the part of the chip that does the computation — the graphics processing unit, or GPU — with several blocks of a fast, specialized component called high-bandwidth memory, or HBM, Sadana said. HBM is often visible when chipmakers hold up their new chips. Micron supplies memory to both Nvidia and AMD, the two leading GPU makers.
Nvidia’s Rubin GPU, which recently entered production, comes with up to 288 gigabytes of next-generation HBM4 memory per chip. HBM is installed in eight visible blocks above and below the processor, and that GPU will be sold as part of single server rack called NVL72, which fittingly combines 72 of those GPUs into a single system. By comparison, smartphones typically come with 8 or 12GB of lower-powered DDR memory.
Nvidia founder and CEO Jensen Huang introduces the Rubin GPU and the Vera CPU as he speaks during Nvidia Live at CES 2026 ahead of the annual Consumer Electronics Show in Las Vegas, Nevada, on Jan. 5, 2026.
Patrick T. Fallon | AFP | Getty Images
But the HBM memory that AI chips need is much more demanding than the RAM used for consumers’ laptops and smartphones. HBM is designed for high-bandwidth specifications required by AI chips, and it’s produced in a complicated process where Micron stacks 12 to 16 layers of memory on a single…
Read More: AI memory is sold out, causing an unprecedented surge in prices