Manufacturing


Semiconductor Type:
HBM Chips



High Bandwidth Memory (HBM) is a vertically stacked DRAM architecture connected via through-silicon vias (TSVs) and placed close to the processor package using 2.5D/3D packaging. HBM enables massive memory bandwidth and lower power consumption, making it the de facto memory standard for AI accelerators, GPUs, and HPC processors.


Role in the Semiconductor Ecosystem

  • Critical enabler for AI training clusters, GPUs, and HPC workloads.
  • Provides 5–10x the bandwidth of DDR5 at significantly lower power per bit.
  • Tightly coupled with advanced packaging (CoWoS, InFO, EMIB) and foundry capacity.
  • Dominant driver of DRAM capex investment in the 2020s.

HBM Generational Roadmap

Generation Bandwidth per Pin Stack Height Vendors Approx. ASP Range Status
HBM2E 3.2–3.6 Gbps Up to 8-high Samsung, SK Hynix, Micron $200–$400 per stack Legacy, used in HPC and some GPUs
HBM3 6.4 Gbps Up to 12-high SK Hynix, Samsung, Micron $400–$800 per stack Mainstream in 2024 AI GPUs (NVIDIA H100, AMD MI300)
HBM3E 8.0–9.2 Gbps 12-high, moving to 16-high SK Hynix (lead), Samsung $800–$1,200 per stack In ramp-up for NVIDIA B200/Blackwell
HBM4 (in R&D) 12–16 Gbps (projected) 16-high+ SK Hynix, Samsung, Micron TBD Target 2026+; EUV + advanced TSV

Vendor Landscape

  • SK Hynix: Market leader in HBM3/3E; main supplier to NVIDIA’s AI GPU line.
  • Samsung: Competing aggressively with HBM3E and preparing HBM4.
  • Micron: Smaller share, but positioned as third supplier for diversification.

Supply Chain Bottlenecks

HBM has become one of the most constrained semiconductor products globally:

  • Packaging: Requires advanced 2.5D/3D packaging (TSMC CoWoS, Samsung I-Cube, Intel EMIB). Capacity is capped and booked out years in advance.
  • Substrates: High-layer-count organic substrates (ABF) are in short supply, limiting production ramp.
  • TSV Yield: Stacking 12–16 dies introduces yield losses; suppliers with better TSV process control gain share.
  • AI-Driven Demand: NVIDIA, AMD, and cloud providers are absorbing nearly all output for AI GPUs and accelerators, leaving little for secondary markets.

Market Outlook

The HBM market was valued at ~$4B in 2023 and is projected to exceed $25B by 2030 (>30% CAGR). HBM demand is directly tied to AI accelerator shipments, making it one of the most strategic choke points in the semiconductor supply chain. By 2030, HBM is expected to account for over 20% of total DRAM capex investment.