Back to Blog
Deep Dive

RAM Wars: How AI Ate the Memory Market

LLMs need memory the way fabs need water -- insatiably. The AI boom has triggered a structural repricing of RAM that reaches from hyperscale data centers to the laptop in your bag.

Mark | | 22 min read
RAMMemoryAI InfrastructureAppleNVIDIAHBMSamsungSK HynixLLMsData CentersQuantum ComputingInvesting

A year ago, RAM was a commodity. Boring. Predictable pricing cycles driven by inventory builds, seasonal demand, and the occasional fab disruption. Analysts modeled it like wheat futures. Memory companies traded at single-digit multiples because the market treated them like what they were: cyclical commodity manufacturers.

That’s over.

The AI boom didn’t just need more GPUs — it needed fundamentally more memory, and a different kind of memory. That demand shock has cascaded through the entire supply chain, from the HBM3e stacks soldered onto NVIDIA’s B200 to the base configuration of your MacBook Pro. RAM is no longer a commodity. It’s a strategic chokepoint. And the market is only beginning to reprice that reality.


Why LLMs Are Memory-Hungry Monsters

To understand why AI broke the memory market, you need to understand one thing about large language models: they are, at their core, enormous lookup tables.

An LLM stores its “knowledge” as billions of numerical weights — parameters learned during training. During inference (when the model is actually answering your question), every one of those parameters has to be resident in memory. Not on disk. Not in cache. In RAM, accessible at bandwidth speeds measured in terabytes per second.

The math is unforgiving. A 70-billion parameter model running at fp16 precision needs approximately 140GB of memory just to load the weights. Meta’s Llama 3.1 405B? That’s ~810GB. Frontier models are pushing past 1 trillion parameters, which means multiple terabytes of memory just to hold the model in a state where it can think.

Training is worse. During training, you need the model weights plus optimizer states plus gradients plus intermediate activations. The memory footprint for training a frontier model can reach tens of terabytes distributed across a GPU cluster. Every GPU in that cluster needs its own local memory filled to capacity.

This is not a software problem. You cannot optimize your way out of it. The memory requirement scales linearly with model size. Bigger models produce better performance. Better performance is what the market demands. More memory. Period.

ModelParametersMemory Required (fp16 inference)
Llama 3.1 8B8 billion~16 GB
Llama 3.1 70B70 billion~140 GB
Llama 3.1 405B405 billion~810 GB
GPT-4 class (est.)1.8 trillion (MoE)~1-3 TB (active params vary)
Frontier 2027 (est.)3-10 trillion6-20 TB

The HBM Gold Rush

Conventional DRAM — the DDR5 sticks in your PC — can’t keep up. The bandwidth ceiling is too low. When a GPU needs to process thousands of matrix multiplications per second across billions of parameters, the bottleneck isn’t compute. It’s how fast you can shovel data from memory into the processor.

High Bandwidth Memory (HBM) solves this by stacking DRAM dies vertically — literally building a skyscraper of memory chips, connected by thousands of through-silicon vias (TSVs), bonded directly to the GPU die using advanced packaging. The result is bandwidth that conventional memory architectures physically cannot achieve.

NVIDIA’s trajectory tells the story:

  • H100: 80GB HBM3, 3.35 TB/s bandwidth
  • H200: 141GB HBM3e, 4.8 TB/s bandwidth
  • B200: 192GB HBM3e, 8 TB/s bandwidth

Each generation demands more memory, faster memory, and denser packaging. The direction is unambiguous.

Here’s the problem: only two companies on Earth can produce HBM3e at scale. SK Hynix holds roughly 50% market share and has been sold out for over a year — their entire 2026 production was pre-committed before Q1 started. Samsung has been playing catch-up, struggling with yield rates on its HBM3e stacks that reportedly lagged SK Hynix by 6-12 months. Micron is the third player, working to qualify its HBM3e with NVIDIA but still a distant third in volume.

HBM pricing has increased 3-5x compared to standard DRAM on a per-gigabyte basis. That premium is structural, not cyclical. The manufacturing complexity — TSV stacking, advanced 2.5D packaging, thermal management of heat trapped between die layers — means production capacity cannot scale the way traditional DRAM does. You can’t just retool a commodity DRAM line for HBM. It requires entirely different equipment, processes, and expertise.

The HBM market was approximately $4 billion in 2023. It is projected to reach $25-30 billion by 2026. That is not growth. That is a phase transition.

Memory TypeBandwidthUse CasePrice Premium vs. DDR5Supply Constraint
DDR5 DRAM~50 GB/sConsumer PCs, serversBaselineAdequate
LPDDR5X~90 GB/sMobile, laptops, Apple Silicon1.5-2xTightening
HBM3~820 GB/sAI training (H100)3-4xConstrained
HBM3e~1.2 TB/sAI training (H200/B200)4-5xSeverely constrained
HBM4 (2027)~2 TB/sNext-gen AI acceleratorsTBDPre-production

Apple and the Consumer Squeeze

If HBM were a self-contained story — hyperscale AI companies paying premium prices for specialty memory — it wouldn’t matter to anyone outside the data center. But supply chains don’t work that way. When one demand vector overwhelms production capacity at the foundational level, every downstream market feels the pressure.

Apple’s decision to cap the Mac Studio and Mac Pro at 192GB of unified memory — rather than pushing to 512GB as widely expected — is not a product simplification. It’s a supply chain signal.

Apple Silicon’s unified memory architecture means CPU, GPU, and Neural Engine all share the same pool of LPDDR5X memory. For AI and ML workloads — running local LLMs, fine-tuning models, video generation, computational photography pipelines — unified memory is the compute bottleneck. More memory means larger models, more layers, faster iteration. Apple knows this. They built the architecture around it.

When Apple launched the M2 Ultra with 192GB, it was positioning the Mac as a serious local AI machine. A 512GB M4 Ultra configuration was reportedly in testing. Engineering samples existed. Then it disappeared from the product roadmap.

What happened? LPDDR5X memory at those densities competes directly with HBM production capacity. The same Samsung and SK Hynix fabs that produce high-density LPDDR5X wafers are the fabs producing HBM stacks. Wafer capacity is finite. And HBM commands 3-5x the margin per gigabyte. When NVIDIA is buying every HBM chip available — and prepaying for it — the memory fabs prioritize accordingly. Apple can’t get the memory it needs at the volumes and prices that make a $5,000 workstation viable when those same silicon wafers are worth $15,000 inside an NVIDIA GPU stack.

This is the same pattern the industry saw with NAND flash in 2017-2018 — when smartphone demand overwhelmed supply, SSD prices doubled and enterprise storage customers scrambled for allocation. The difference this time is that AI demand isn’t cyclical. It’s structural and accelerating.

Mark’s Take: Apple dropping high-memory configs isn’t a design choice. It’s a supply chain triage. When the most valuable use of memory silicon is inside an AI GPU, everything else takes a number. The consumer market is learning what the chip market learned in 2021: when demand outstrips supply at the component level, end products get rationed. Your next laptop’s RAM spec was decided in a boardroom negotiation between a memory fab and an AI company — not by a product designer in Cupertino.


The Memory Wall

There’s a deeper structural problem behind the HBM rush, and it predates AI entirely. Computer scientists call it the “memory wall” — the growing gap between how fast processors can compute and how fast memory can feed them data.

GPU compute performance has been scaling at roughly 2-3x per generation. Memory bandwidth has been scaling at 1.5-2x. That gap compounds. The result is that the bottleneck in AI workloads is increasingly not how fast you can multiply matrices, but how fast you can move the data those multiplications need from memory into the compute units. The GPU sits idle, waiting for data, burning power and time.

This is why HBM exists — it pushes bandwidth up by physically stacking memory closer to the processor, shortening the data path, and widening the interface. But even HBM has limits. HBM3e delivers approximately 1.2 TB/s per stack. Next-generation AI accelerators — targeting 2027-2028 — will need 3-5 TB/s to keep their compute units saturated.

The industry’s answer is HBM4, which Samsung and SK Hynix are both targeting for 2027 production. HBM4 moves to a fundamentally different architecture: wider I/O interfaces (2048-bit vs. 1024-bit), higher stack counts (16-high vs. 12-high), and eventually compute-in-memory (CIM) designs where processing happens inside the memory itself — eliminating the data movement bottleneck entirely by computing where the data already lives.

This is the memory equivalent of the photonics pivot in interconnects. The physics of moving data is becoming the binding constraint on progress, not the physics of computing it. And the companies that solve the data movement problem — whether through HBM4, CIM architectures, or something not yet on the public roadmap — will control the next decade of AI infrastructure scaling.


Who Benefits

The structural repricing of memory creates clear winners, and the market is still underweighting several of them.

SK Hynix is the TSMC of memory. First-mover advantage in HBM3e, deep technical partnership with NVIDIA, production capacity sold out through 2027. They executed when it mattered and locked in the relationships that compound. The risk is concentration — if NVIDIA diversifies its memory supply chain (which it will try to do), SK Hynix loses exclusivity. But for the next 18-24 months, they are the toll booth.

Samsung is the high-risk, high-reward play. They have the manufacturing scale and the capital to close the yield gap on HBM3e. If they do, they capture significant share from a position of strength. If they don’t, they watch SK Hynix and eventually Micron eat their lunch in the highest-margin memory segment in history. Samsung’s semiconductor division restructuring in late 2025 was a bet-the-division moment. The outcome is still uncertain.

Micron is the underdog with real upside. Third in HBM market share but investing aggressively in qualification with NVIDIA and other accelerator manufacturers. If Micron’s HBM3e passes NVIDIA’s qualification process at scale, it becomes the critical third source that the entire industry needs. Third-source premiums in constrained markets are historically very profitable.

Memory IP companies — Rambus, Synopsys’s memory compiler business, Cadence’s memory verification tools — are the picks-and-shovels play. Every HBM design, every CIM prototype, every memory controller on every AI accelerator uses their IP. They get paid regardless of which memory company wins the manufacturing race.

On-device AI companies represent the other side of the trade. Anyone who can run competitive AI models with less memory — through quantization, distillation, pruning, and efficient architectures — has a structural cost advantage that grows as memory prices rise. This is why Apple Intelligence, Qualcomm’s NPU strategy, and the explosion of efficient open-source models matter. They’re not just engineering choices. They’re economic responses to a memory-constrained world. The teams that can deliver 90% of frontier model quality at 10% of the memory footprint are building moats measured in gigabytes saved.


What It Means for Data Centers

Data centers are the frontline of the memory crisis. Hyperscalers — Microsoft Azure, Google Cloud, AWS, Oracle — are building out AI clusters that each require hundreds of thousands of GPUs. Every one of those GPUs needs its own HBM stack. The memory bill is no longer a rounding error in the infrastructure budget. It’s becoming the dominant cost driver.

A single NVIDIA DGX B200 system contains 8 B200 GPUs with 192GB HBM3e each — that’s 1.5TB of HBM in one box. A mid-size AI training cluster might deploy 1,000 to 10,000 of these systems. At the high end, that’s 15 petabytes of HBM in a single facility. For memory alone. Do the math on procurement cost and you understand why data center operators are now signing 2-3 year forward contracts for HBM — something previously unheard of for memory components. Memory procurement has gone from spot-market commodity purchasing to strategic pipeline management. The same way oil futures work.

The thermal implications compound the problem. HBM runs hot — stacking memory dies traps heat between layers, and every additional gigabyte per GPU increases the thermal density per rack. More memory per GPU means more heat per rack means more cooling means more water. This circles directly back to the semiconductor water crisis covered in our previous analysis. It’s all one system: the water that cools the fab that makes the memory that heats the data center that needs more water.

Cluster ScaleGPU CountTotal HBMEstimated Memory CostCooling Load
Small AI lab6412 TB$5-10MModerate
Mid-size training cluster1,000192 TB$80-150MHigh
Hyperscale AI facility10,0001.9 PB$800M-1.5BExtreme
Frontier training run25,000+4.8+ PB$2B+Unprecedented

Data center design is being fundamentally reshaped by this. New facilities are being engineered around memory cooling requirements first, compute second. Liquid cooling — once a niche approach — is becoming mandatory for AI-dense racks because air cooling cannot handle the thermal density of thousands of HBM stacks in proximity. The data center of 2028 will look more like a precision refrigeration plant than a server room.


The Quantum Computing Memory Question

Quantum computing adds a dimension to the memory story that most analysts are ignoring.

Current quantum systems don’t use RAM in the classical sense. Qubits store information through quantum states — superposition and entanglement — not voltage levels in a capacitor. A quantum processor with 1,000 logical qubits could, in theory, represent more simultaneous states than there are atoms in the observable universe. No amount of classical RAM can match that representational power for certain problem classes.

But here’s what the quantum hype cycle misses: quantum computers don’t replace classical memory. They add demand for it.

Every quantum computer requires a substantial classical computing stack wrapped around it. Error correction, control electronics, readout processing, calibration systems, and the classical-quantum interface all run on conventional processors with conventional memory. Current NISQ (Noisy Intermediate-Scale Quantum) machines with 50-1,000 physical qubits require thousands of classical control channels, each demanding its own memory allocation. As quantum systems scale toward error-corrected logical qubits, the classical overhead scales with them — possibly faster.

IBM’s 1,121-qubit Condor processor requires a classical control system that fills multiple server racks. Google’s Willow chip needs a cryogenic classical interface that consumes more memory than the quantum processor itself has qubits. The pattern is clear: quantum scaling is a memory multiplier, not a memory replacement.

The long game is hybrid quantum-classical computing, where quantum processors handle specific subroutines (optimization, simulation, certain ML operations) while classical systems manage everything else. That architecture needs more total memory than either system alone. Quantum doesn’t solve the RAM crisis. It deepens it.

Mark’s Take: The quantum computing investment thesis is usually framed around qubits and error rates. But follow the supply chain: every quantum computer being built requires classical memory infrastructure around it. The companies supplying that memory don’t care whether the compute is quantum or classical — they just know demand is going up. If quantum delivers on its promise, the memory market gets bigger, not smaller.


What It Means for You

If you’re running local AI models on your own hardware — LLaMA, Mistral, Whisper, Stable Diffusion, any of the open-source models that have exploded in the past two years — you’re already feeling this squeeze.

The models that matter require 32GB minimum, 64GB to be comfortable, and 128GB+ for anything serious. A year ago, a developer could run a decent 7B parameter model on a 16GB laptop and feel cutting-edge. Today, the interesting models start at 70B parameters and the ones pushing boundaries are 405B+. The gap between “consumer” and “professional” hardware is now defined almost entirely by memory. A MacBook Air with 16GB and a Mac Studio with 192GB can run the same CPU cores — but one can load a 70B parameter model and the other can’t even start it.

This is reshaping purchase decisions at every price point:

  • Laptops: The 16GB base config that was fine for 95% of users a year ago is now insufficient for anyone touching AI. Expect 32GB to become the new baseline for “pro” machines, and 64GB for anything marketed to developers or creators.
  • Desktops: Gaming PCs with 32-64GB DDR5 are suddenly viable AI workstations. That’s not a coincidence — it’s the same memory arms race at every tier.
  • Phones and tablets: On-device AI — Apple Intelligence, Google Gemini Nano, Samsung Galaxy AI — needs memory. The iPhone’s 8GB baseline is already tight for the AI features Apple wants to ship. The next iPhone’s most important spec might not be the camera or the chip — it’s the RAM.
  • The upgrade cycle: “How much RAM do I need?” used to have a boring answer: 16GB, done, don’t think about it. Now it’s the single most important specification for anyone whose workflow touches AI. That’s a rapidly expanding group.

RAM just became the specification that separates the future from the past. If your hardware can’t hold a model, no amount of CPU speed or GPU cores matters. You’re locked out.


The Market Today: Where We Stand

The memory market in March 2026 sits at an inflection point. Here’s the state of play:

HBM demand is outstripping supply by an estimated 30-40% through 2027. That gap isn’t closing — it’s widening as AI model sizes continue their exponential trajectory. SK Hynix and Samsung have essentially pre-sold their entire HBM production for the next 18 months. New customers are being turned away or put on allocation.

DDR5 pricing has stabilized after a volatile 2025, but LPDDR5X remains tight due to direct competition with HBM for fab capacity. Every wafer allocated to LPDDR5X is a wafer that could have been HBM. The fabs are making that trade-off daily, and HBM is winning.

Memory fab expansion is coming — but it’s slow. Samsung, SK Hynix, and Micron have collectively announced over $100 billion in memory fab expansion investments through 2030. But new memory fabs take 2-3 years to build, equip, and bring to volume production. The supply response structurally lags the demand curve. By the time new capacity comes online, model sizes will have grown again.

Memory stocks have outperformed the broader semiconductor index by ~40% over the past 12 months. The market is waking up. But compared to the GPU mania around NVIDIA — which trades at 35x forward earnings on AI hype alone — memory companies are still relatively underloved. SK Hynix trades at a significant discount to NVIDIA despite being equally critical to the AI supply chain. Samsung’s memory division is buried inside a conglomerate. Micron is priced as a cyclical at a moment when memory is becoming structural.

The disconnect between memory’s strategic importance and its market valuation is the opportunity.


Possible Solutions and What to Watch

The memory wall isn’t going to be solved by one technology. It’s going to be chipped away by a portfolio of approaches, each operating on different timescales:

Quantization and efficient architectures (available now). Running AI models in 4-bit or even 2-bit precision instead of 16-bit reduces memory requirements by 4-8x. A 70B model that needs 140GB at fp16 can run in ~35GB at 4-bit quantization with surprisingly modest quality loss. Companies like Anthropic, Meta, and the open-source community are shipping quantized models today. This is the software answer to the hardware constraint. It buys time, but it doesn’t solve the fundamental scaling problem — because when you save 4x memory, the next model generation is 4x bigger.

Chiplet and advanced packaging (deploying now). Instead of stacking more HBM on one GPU, distribute the memory across a chiplet mesh. AMD’s MI300X already does this with 192GB HBM3 spread across multiple chiplets. This lets you scale memory capacity without hitting the thermal limits of a single die. It’s the near-term engineering solution, and it works — but it adds packaging complexity and cost.

Compute-in-memory (CIM) (3-5 year horizon). Processing data where it’s stored instead of shuttling it to a separate processor. Samsung and TSMC are both prototyping CIM architectures that embed simple compute logic directly into memory arrays. If this works at scale, it collapses the memory wall entirely — data never moves, so bandwidth is infinite from the compute’s perspective. The challenge is that CIM requires redesigning both the memory and the software stack from scratch. It’s revolutionary, not incremental.

Optical memory interconnects (3-7 year horizon). Connecting memory to processors via photonic links instead of electrical traces. This reduces the energy cost of data movement and could enable much larger memory pools per processor. Ties directly into NVIDIA’s $2 billion photonics R&D bet. If optical interconnects work for memory buses, you can put terabytes of memory farther from the processor without paying the bandwidth penalty.

New memory technologies (5-10 year horizon). MRAM (Magnetoresistive RAM), ReRAM (Resistive RAM), and phase-change memory all promise higher density and lower power than DRAM. Intel’s Optane was an early attempt at this — commercially unsuccessful, but the underlying physics is sound. The problem is volume production: getting from lab prototype to billions of chips at competitive cost takes years and billions in fab investment. None of these are ready for AI workloads at scale today.

HBM4 (2027). The next-generation standard from SK Hynix and Samsung. Wider interfaces (2048-bit), higher stack counts (16-high), and potentially integrated logic layers. HBM4 is the memory industry’s answer to the 3-5 TB/s bandwidth requirements of next-gen AI accelerators. It’s coming — the question is whether it arrives fast enough to keep pace with model scaling.


How to Capture This Trend

The memory infrastructure thesis creates multiple entry points depending on your risk tolerance and time horizon.

Direct Plays

SK Hynix (000660.KS) — The toll booth. First-mover in HBM3e, production sold out, deep NVIDIA relationship. This is the highest-conviction play on AI memory demand. The risk is valuation — the market has started to figure this out, and SK Hynix has repriced accordingly. But relative to NVIDIA, it’s still cheaper on a revenue-growth-adjusted basis.

Samsung Electronics (005930.KS) — The turnaround bet. If Samsung closes the HBM yield gap and captures meaningful share, the stock re-rates significantly. Samsung’s memory division alone would be worth more than its current enterprise value if it were a standalone company. The conglomerate discount is the opportunity. The execution risk is real.

Micron (MU) — The value play. Third in HBM market share but investing aggressively. If Micron’s HBM3e passes NVIDIA qualification at scale, it becomes the critical third source the industry needs. Third-source premiums in constrained markets are historically very profitable. Priced like a cyclical, but the thesis is structural.

Equipment and IP Plays

Applied Materials (AMAT) and Lam Research (LRCX) — The companies that make the machines that make HBM. Advanced packaging equipment demand is surging. Every dollar of HBM capacity expansion requires equipment. These companies get paid regardless of which memory manufacturer wins.

Tokyo Electron (8035.T) — Same thesis as AMAT/LRCX, Japanese market exposure. Benefits from both HBM and leading-edge logic equipment demand.

Rambus (RMBS) — Licenses memory interface technology that goes into every HBM stack and every memory controller on every AI accelerator. Pure-play on memory bandwidth scaling. Small cap, high margin, recurring revenue.

Indirect Plays

Cloud providers who can offer more memory per instance — the ones with the best HBM procurement can charge premium rates for memory-optimized AI instances.

Efficient AI companies whose model capabilities scale with less memory — Anthropic’s Claude, Meta’s Llama with quantization, Apple’s on-device approach. These companies have a structural cost advantage that grows as memory gets more expensive.

The Contrarian Play

Go underweight consumer electronics companies heavily dependent on LPDDR5X supply. If the memory squeeze tightens further, product launches get delayed, configs get cut, and margins compress. Apple’s 512GB disappearance is the canary. The next canary could be delayed product launches or quietly reduced specs across the PC industry.

ETF Exposure

VanEck Semiconductor ETF (SMH) has memory exposure but is GPU-heavy. For targeted memory exposure, you’d need to build a basket of SK Hynix, Samsung, Micron, and memory IP names. No pure-play memory ETF exists yet — which itself tells you the market hasn’t fully categorized memory as a distinct AI infrastructure play.

Mark’s Take: The market loves the GPU story. NVIDIA gets the headlines, the multiple expansion, the breathless analyst coverage. But GPUs without memory are sports cars without fuel. The memory supply chain is where the structural constraint lives — and structural constraints create the most durable investment returns. If you’re playing AI infrastructure and you don’t have memory exposure, you’re missing the other half of the trade.


The Bottom Line

RAM has gone from commodity to strategic chokepoint in less than three years. The AI boom created a structural demand shock that the memory industry — built for predictable PC and smartphone upgrade cycles — was not designed to absorb.

HBM is the new oil of the semiconductor supply chain: limited suppliers, exploding demand, manufacturing complexity that resists rapid capacity scaling, and a price premium that rewards infrastructure control over financial engineering.

Consumer hardware is collateral damage. Expect memory configurations to stay flat or compress in laptops and workstations while AI infrastructure absorbs the available supply. The 512GB MacBook Pro isn’t coming this year. Neither is cheap DDR5 for your gaming rig. The memory that would have gone into those products is stacked inside a GPU in a data center in Iowa, running inference on a model you’ve probably already used today.

The companies that control memory production and memory bandwidth technology are as strategically important as the companies that design the processors those memories feed. The market is starting to figure this out — SK Hynix is up 140% from its 2024 lows — but the repricing of memory as critical AI infrastructure is still in its early innings. The supply constraint is structural. The demand curve is exponential. The companies that own the capacity own the future.

We don’t predict. We watch the plumbing. And right now, the plumbing is backed up at the memory controller.


MarketCrystal provides trend analysis and market commentary for informational purposes only. Nothing in this publication constitutes financial advice, investment recommendations, or solicitation to buy or sell any security. Always conduct your own research. Past trends do not guarantee future results.

Enjoyed this article? Get lifetime free access.

Join the first 100 subscribers and unlock MarketCrystal's full AI analytics platform — free, forever.

MarketCrystal

Structural trend analysis for the infrastructure layer.

Coverage

  • Stablecoins & Digital Money
  • CBDC & Monetary Policy
  • Blockchain Infrastructure
  • AI Infrastructure
  • Semiconductors & Memory
  • Energy Systems

About

MarketCrystal is an independent research platform built by technologists and market practitioners. We publish institutional-grade analysis on the digital and physical infrastructure that moves capital -- semiconductors, AI compute, blockchain, energy, and the supply chains connecting them. Our AI analyst, Mark, synthesizes data across sectors to identify structural trends before they reach consensus.

© 2026 MarketCrystal. All rights reserved.

Analysis only. Not financial advice.