Memory Chips: The Silent Force Driving AI’s Next Leap

Prefer to listen instead? Here’s the podcast version of this article.

The artificial intelligence revolution is often measured in terms of algorithms and compute power, but there’s another critical factor shaping its trajectory: memory chips. As AI models grow larger and more complex, their need for rapid access to vast amounts of data has placed unprecedented demands on memory technology. High-Bandwidth Memory (HBM) and emerging compute-in-memory architectures are stepping into the spotlight, not just as supporting components, but as foundational technologies driving the next wave of AI innovation. Understanding this shift is essential for businesses, technologists, and policymakers alike, as memory is poised to redefine the limits of what AI systems can achieve.

 

Overcoming the “Memory Wall” in AI Infrastructure

AI models, especially LLMs, demand massive memory bandwidth. As reported by the Financial Times, AI compute has exploded—while DRAM bandwidth has barely budged, leading to a “memory wall” that throttles performance. HBM changes that by delivering significantly higher bandwidth via stacked memory architectures, lowering latency and boosting efficiency. [Financial Times]

 

HBM: From Niche to Center Stage

HBM is no longer just memory—it’s mission-critical. HBM (1 → 4) has evolved rapidly, offering multi-TB/s bandwidth and dense stacking. The JEDEC standard HBM4, released in April 2025, supports up to 2 TB/s per stack and 64 GB capacity—a leap from earlier generations.

 

Industry Shakeup: SK Hynix vs. Samsung

SK Hynix has surged ahead of Samsung in DRAM revenue, thanks to its prowess in HBM. By Q1 2025, HBM accounted for over 40% of Hynix’s DRAM revenue, up from just 5% in late 2022. SB experts attribute this to HBM’s pricing leverage—strong margins and tailored, long-term contracts—making memory companies more strategic players in AI infrastructure. [Financial Times]

 

The Next Frontier: HBM4 with Logic Die & Hybrid Bonding

HBM4 isn’t just faster—it’s smarter. With logic dies powered by advanced processors (via TSMC or Samsung’s foundry) and hybrid bonding techniques, HBM4 promises even greater bandwidth, efficiency, and task-specific tuning. This positions memory chips as not just passive storage but optimized co-processors.

 

Beyond HBM: Compute-in-Memory & Memory-Centric Computing

Real innovation lies even deeper. Compute-in-memory (PIM) and memory-centric computing integrate compute capabilities directly into memory chips.

 

  • A recent arXiv paper introduces PIM-AI, a DDR5/LPDDR5 PIM architecture that slashes total cost of ownership (TCO) and energy per query—up to 6.94× lower energy than GPUs in cloud LLM inference scenarios.

  • Onur Mutlu’s work on memory-centric computing explores architectures enabling computation in or near memory, delivering orders-of-magnitude gains in performance and energy across workloads.

 

Conclusion

The future of AI will not be determined by algorithms alone—it will be shaped by the memory systems that enable those algorithms to run at scale. From High-Bandwidth Memory (HBM) driving efficiency in today’s data centers to compute-in-memory architectures promising revolutionary gains in performance and energy use, memory chips are evolving from silent enablers into central players in the AI ecosystem.

WEBINAR

INTELLIGENT IMMERSION:

How AI Empowers AR & VR for Business

Wednesday, June 19, 2024

12:00 PM ET •  9:00 AM PT