The Conversation Is Shifting from GPUs to Memory
For the past several years, the narrative around AI infrastructure costs has been dominated by a single topic: Nvidia GPUs. The scarcity, pricing, and allocation of graphics processing units have driven headlines, investment decisions, and corporate strategy across the technology industry. But a quieter shift is underway in how the industry thinks about AI infrastructure economics. Increasingly, memory, not processing power, is emerging as the binding constraint on AI system performance and cost.
The dynamic makes intuitive sense when you examine how modern AI models actually operate. A large language model does not simply compute answers. It must hold vast amounts of data in active memory, accessible at extremely high speeds, to process each request. The model's weights, the numerical parameters that encode its knowledge and capabilities, must be loaded into memory before inference can begin. For frontier models with hundreds of billions or even trillions of parameters, the memory required to hold these weights dwarfs what conventional computing systems were designed to provide.
High Bandwidth Memory: The Critical Component
The specific type of memory that has become central to AI infrastructure is High Bandwidth Memory, known as HBM. Unlike the standard DRAM found in consumer computers, HBM stacks multiple layers of memory chips vertically and connects them with an extremely wide data bus, enabling data transfer rates that are orders of magnitude faster than conventional memory. This speed is essential because AI accelerators like Nvidia's H100 and H200 GPUs can process data far faster than standard memory can deliver it. Without HBM, these processors would spend most of their time waiting for data, rendering their computational capabilities largely useless.
HBM is physically bonded to the AI accelerator using advanced packaging techniques, creating an integrated module where memory and processing are tightly coupled. This integration provides the bandwidth needed for AI workloads but also creates a supply chain dependency: every AI accelerator shipped requires a corresponding allocation of HBM, and the production capacity for HBM is concentrated among just three manufacturers globally.
A Three-Company Oligopoly
The global supply of HBM is controlled by three companies: SK hynix, Samsung, and Micron. SK hynix, the South Korean semiconductor manufacturer, currently dominates the market and is Nvidia's primary HBM supplier. Samsung, despite being the world's largest memory chip company by overall revenue, has struggled with yield issues in its HBM production and has lost significant market share to SK hynix in this critical segment. Micron, the American memory manufacturer, has been gaining ground with competitive HBM products but operates at a smaller scale than its Korean rivals.
This concentrated supply structure creates significant pricing power for HBM manufacturers and vulnerability for AI infrastructure companies. When demand outpaces supply, as it has consistently over the past two years, prices rise and allocation becomes a strategic negotiation rather than a straightforward procurement process. Companies building AI data centers must secure HBM commitments well in advance, often signing long-term supply agreements at premium prices to ensure they can obtain the memory needed for their planned deployments.
The economics are striking. HBM can represent 30 to 40 percent of the total cost of an AI accelerator module, a proportion that has been growing as HBM prices increase faster than the broader semiconductor market. For a company deploying thousands of AI accelerators in a new data center, the memory bill alone can run into hundreds of millions of dollars.
Why Demand Keeps Growing
Several trends are converging to intensify demand for HBM and AI-grade memory more broadly. The most obvious is the continued growth in model sizes. Each new generation of frontier AI models tends to be significantly larger than its predecessor, requiring proportionally more memory to store its parameters. But model size is only part of the equation.
Inference demand is arguably a more significant driver of memory consumption than training. While training a model is a one-time (or periodic) process that requires massive computational resources for a finite period, inference, the process of actually running the model to respond to user requests, is continuous and scales with user adoption. Every chat interaction, every code completion, every image generation request requires loading model weights into memory and keeping them there for the duration of processing.
As AI applications proliferate and user adoption grows, the aggregate inference demand across the industry is growing exponentially. Companies are deploying models in customer service, software development, content creation, data analysis, and hundreds of other applications, each generating continuous memory demand. The total memory required to serve all of these workloads simultaneously now represents a significant fraction of global HBM production capacity.
Context window expansion is another factor. Models like Anthropic's Claude and Google's Gemini now offer context windows of one million tokens or more, meaning they can process vast amounts of input text in a single request. Handling these large contexts requires storing attention states and intermediate computations in memory throughout the processing pipeline, adding to per-request memory consumption.
The Ripple Effects on Infrastructure Planning
Memory constraints are beginning to influence AI infrastructure decisions in ways that would have seemed unlikely even two years ago. Data center architects are designing systems with memory provisioning as a primary constraint rather than an afterthought. Cloud providers are creating memory-optimized instance types specifically for AI inference workloads. And hardware companies are exploring novel memory technologies that could provide higher capacity or bandwidth at lower costs.
The memory challenge also affects model development decisions. Some AI labs are investing heavily in techniques to reduce the memory footprint of their models without sacrificing capability, including quantization, which reduces the numerical precision of model weights, and mixture-of-experts architectures, which activate only a subset of a model's parameters for each request. These techniques are not just academic exercises. They are direct responses to the practical constraint that memory imposes on deployment economics.
For the broader AI ecosystem, the shift in attention from GPUs to memory represents a maturation of understanding about what actually determines the cost and feasibility of AI deployment at scale. The GPU shortage narrative, while not entirely resolved, has been partially addressed by increased production capacity and the entrance of competitors like AMD and custom silicon from major cloud providers. Memory, by contrast, faces longer lead times for capacity expansion and fewer competitive alternatives, making it a more persistent and structurally challenging bottleneck.
What Comes Next
The memory companies are responding to demand with ambitious capacity expansion plans. SK hynix is building new production facilities and ramping output of its latest HBM3E products. Samsung is working to resolve its yield issues and regain competitive standing. Micron is investing in expanded HBM production in both the United States and Japan. But semiconductor manufacturing capacity takes years to build, and the gap between current supply and projected demand suggests that memory will remain a constraining factor in AI infrastructure for the foreseeable future.
Emerging technologies like Compute Express Link, which allows systems to share memory pools across multiple processors, and new memory architectures being developed in research labs could eventually ease the constraint. But these solutions are years from commercial deployment at scale. In the meantime, the AI industry is learning that the infrastructure challenge is not about any single component but about the complex interplay of processors, memory, networking, power, and cooling that together determine what is possible and at what cost.
This article is based on reporting by TechCrunch. Read the original article.




