SemiconductorX > Sectors > AI & ML


Datacenter / HPC
Semiconductors



The datacenter and high-performance computing sector is the largest single consumer of semiconductor capital expenditure in the world when measured by total system value, and the sector undergoing the most rapid architectural transformation in its history. Two structural forces are driving simultaneous change. First, the AI infrastructure buildout is layering a new high-density compute tier - GPU and accelerator clusters with 10-120kW per rack power density - on top of the existing general-purpose cloud compute base. Second, the x86 server CPU architecture that has dominated datacenter compute since the 1990s is losing ground to ARM-based designs at the hyperscaler level, driven by power efficiency economics and the vertical integration strategies of AWS (Graviton), Microsoft (Cobalt), Google (Axion), and Ampere Computing. Both forces are reshaping semiconductor supply chains simultaneously and in ways that interact with each other.

The SX editorial boundary between this page and the AI & ML sector page is deliberate. The AI sector page owns the accelerator supply chain - GPU dies, HBM, CoWoS, and custom ASIC programs at hyperscalers. This page owns the general-purpose datacenter semiconductor stack: server CPUs, DDR5 server DRAM, NVMe storage controllers, Ethernet switch ASICs, data processing units (DPUs), power delivery silicon, and the infrastructure semiconductor layer that supports both conventional cloud workloads and AI cluster deployments. The two sectors share physical infrastructure but have distinct semiconductor supply chains and distinct supply chain risk profiles.

Related Coverage: CPUs | Memory & Storage | FPGAs | ASICs | GPU & CPU Boards | AI & ML Sector | Bottleneck Atlas

Semiconductor Device Map — Datacenter / HPC

Function Device types Key suppliers Foundry / node Supply chain status
Server CPU (x86) High core-count server CPUs with large L3 cache, PCIe 5.0, DDR5 memory controller, CXL support; dual-socket configurations for enterprise workloads Intel (Xeon Scalable - Sapphire Rapids, Emerald Rapids, Granite Rapids); AMD (EPYC Genoa, Bergamo, Turin) Intel 7 / Intel 3 (Intel internal); TSMC N5 / N4 (AMD EPYC Turin); TSMC N4 (AMD EPYC Genoa) Competitive - AMD EPYC gaining share at hyperscalers due to performance-per-watt advantage; Intel defending position with Xeon while transitioning internal process; supply broadly adequate; AI cluster CPU demand (host CPUs for GPU nodes) adds secondary demand vector
Server CPU (ARM) Custom ARM Neoverse-based server SoCs; in-house hyperscaler ARM CPUs; high efficiency-per-watt designs for general cloud compute AWS (Graviton4 - custom ARM Neoverse V2); Google (Axion - custom ARM Neoverse V2); Microsoft (Cobalt 100 - custom ARM Neoverse N2); Ampere (AmpereOne - up to 192 cores); Qualcomm (QDFN server ARM SoC) TSMC N5/N4 (Graviton4, Axion, Cobalt); TSMC N5 (AmpereOne) Growing rapidly - ARM server CPU share at hyperscalers estimated 30-40% of new CPU deployments at AWS and Google by 2025; structural cost and power efficiency advantage over x86 at cloud-native workloads; competes for TSMC N5 allocation with AI programs
Server DRAM DDR5 RDIMMs and LRDIMMs for server memory; LPDDR5X for low-power server variants; CXL memory expansion modules; ECC memory for mission-critical enterprise workloads Samsung (dominant server DRAM share); SK Hynix; Micron (largest US-based DRAM supplier) Samsung Pyeongtaek and Hwaseong internal DRAM fabs; SK Hynix Icheon and Cheongju; Micron Boise and Hiroshima (Manassas VA for ECC DRAM) Cyclical - DDR5 transition from DDR4 creating mixed installed base; DRAM pricing highly cyclical; AI server DDR5 demand (host memory for GPU nodes) growing independently of general server DDR5; CXL memory at early commercial stage with Samsung and SK Hynix leading
NVMe storage controller PCIe 5.0 NVMe SSD controllers; enterprise NVMe controllers with power loss protection; computational storage drive (CSD) controllers with embedded compute; RAID-on-chip ASICs for storage arrays Marvell (88SS9000 series enterprise NVMe), Phison (E26 PCIe 5.0), Samsung (in-house controller for Samsung PM series), Western Digital (in-house), Broadcom (SAS/SATA RAID controllers), Microchip (Flashtec series) TSMC N7/N5 for high-end enterprise NVMe controllers; TSMC N12/N16 for mid-range; Broadcom RAID at N7 Adequate - enterprise NVMe controller supply recovered from 2021-2023 shortage; PCIe 5.0 transition adding new design cycle; NAND flash supply cycles independently of controller supply
Ethernet switch ASIC High-radix Ethernet switch ASICs (51.2 Tbps and above); data center top-of-rack (ToR) switch ASICs; spine and core switch ASICs; smart NIC ASICs with offload engines Broadcom (Tomahawk 5 - 51.2Tbps, Jericho3-AI for AI cluster fabric); Marvell (Teralynx 10, Prestera); Intel (Tofino - discontinued 2023); Cisco (Silicon One Q200); Innovium (acquired by Marvell) TSMC N5/N7 for leading-edge switch ASICs (Tomahawk 5 at N5); TSMC N7 for mid-range Marvell Broadcom dominant - Tomahawk switch ASICs in essentially every hyperscaler Ethernet fabric; AI cluster transition from InfiniBand to Ultra Ethernet driving Jericho3-AI demand surge; Intel Tofino discontinuation leaves programmable switch ASIC market to P4-enabled alternatives
Data processing unit (DPU) SmartNIC / DPU SoCs offloading networking, storage, and security processing from server CPU; PCIe-attached DPU cards; CXL-attached compute offload NVIDIA (BlueField-3 DPU - dominant hyperscaler DPU); Marvell (OCTEON 10 DPU); Broadcom (Stingray PS250); AMD (Pensando Pollara DPU, acquired 2022); Intel (IPU E2100 - Mount Evans ASIC for Google) TSMC N5 (BlueField-3); TSMC N5 (OCTEON 10); TSMC N7 (Stingray) Growing - DPU adoption at hyperscalers accelerating as CPU core efficiency gains slow and network/storage offload value increases; NVIDIA BlueField-3 in widespread deployment at AWS, Azure, Google; GPU cluster DPU demand adds to general hyperscaler DPU demand
Power delivery (GaN / VRM) GaN-based server PSU ICs (48V rack power conversion); voltage regulator modules (VRM) for CPU and GPU power delivery; digital PWM controllers; hot-swap controllers; rack-level power distribution ICs Navitas Semiconductor (GaNFast for server PSU), Infineon (CoolGaN), TI (server PMIC and VRM controllers), Monolithic Power Systems (MPS - AI server PMICs), Vicor (48V point-of-load modules), Renesas (RAA series digital PWM) GaN-on-silicon at TSMC (Navitas, Infineon); server PMIC at 40nm-130nm mixed-signal nodes; VRM controllers at mature nodes Tightening - AI cluster power density (10-120kW per rack for GB200 NVL72) driving GaN PSU and VRM demand far above historical server baselines; OCP (Open Compute Project) 48V rack power standard accelerating GaN PSU adoption; power delivery emerging as secondary AI cluster constraint
FPGA (datacenter acceleration) High-end datacenter FPGAs for network acceleration, financial compute, genomics, and database offload; FPGA-based SmartNICs; adaptive compute acceleration platforms (ACAPs) AMD/Xilinx (Virtex UltraScale+, Alveo U series, Versal ACAP); Intel/Altera (Stratix 10, Agilex - Agilex 9 for datacenter); Microchip (PolarFire for lower-power applications) TSMC N16 (Virtex UltraScale+); TSMC N7 (Alveo U55C, Versal); Intel 10ESF (Stratix 10); TSMC N7 (Agilex 7); TSMC N5 (Agilex 9) Stable - datacenter FPGA market mature; GPU and custom ASIC displacement reducing FPGA share in inference offload; FPGA retaining position in low-latency network acceleration, real-time processing, and prototyping applications where reconfigurability justifies cost premium
Optical interconnect Coherent optical transceiver ASICs (400G, 800G, 1.6T); pluggable optical modules (QSFP-DD, OSFP); co-packaged optics (CPO) DSP silicon; silicon photonics PICs Coherent Corp (optical transceiver DSP); Marvell (PAM4 DSP for 800G transceivers); Broadcom (CPO silicon photonics); Inphi/Marvell (acquired); II-VI/Coherent (module manufacturing); Lumentum (laser sources); Intel (silicon photonics - Intel Integrated Photonics Solutions) DSP ASICs at TSMC N7/N5; silicon photonics PICs at specialty photonics foundries (GlobalFoundries, IME Singapore); laser components at III-V specialty fabs Constrained - 800G transceiver supply tight relative to AI cluster buildout demand; 1.6T transceivers entering production 2025-2026; co-packaged optics (CPO) integration with switch ASICs is the long-term architecture but requires new supply chain for integrated photonic-electronic packages; III-V laser component supply a secondary constraint
Security silicon Hardware roots of trust (RoT); TPM 2.0 security controllers; crypto acceleration ASICs; confidential computing silicon (Intel TDX, AMD SEV-SNP); hardware security modules (HSM) Nuvoton (NPCT TPM 2.0), Infineon (SLB TPM), Microsoft (Pluton security processor in AMD and Intel CPUs), Google (Titan security chip for GCP), AWS (Nitro security controller), AMD (PSP - Platform Security Processor integrated in EPYC) Discrete TPM at 40nm-90nm specialty nodes; integrated security processors in CPU die at leading-edge node of host CPU; HSM at various foundries Adequate - TPM and RoT supply normalized post-2022 shortage; confidential computing silicon integrated into CPU die at AMD and Intel reducing discrete security IC dependency; sovereign cloud programs driving demand for hardware attestation silicon

The x86-to-ARM Server CPU Transition

The most structurally significant long-term shift in datacenter compute silicon is the erosion of x86 dominance by ARM-based server CPUs. This transition is not driven by a single technical breakthrough - it is driven by the compounding of three simultaneous advantages that ARM designs hold over x86 at cloud-native workloads. First, power efficiency: ARM Neoverse cores achieve comparable integer compute performance to x86 cores at significantly lower power draw, which matters enormously at hyperscaler scale where electricity cost dominates total cost of ownership. Second, die cost: ARM server SoCs designed without the legacy x86 instruction set compatibility overhead can be implemented more efficiently at a given node, reducing die area and cost per core. Third, vertical integration: the hyperscalers designing their own ARM CPUs (AWS Graviton, Google Axion, Microsoft Cobalt) capture the margin previously paid to Intel and AMD, control their own performance and power roadmaps, and can differentiate their cloud offerings based on custom silicon that competitors cannot replicate.

AWS Graviton is the most mature and commercially significant hyperscaler ARM CPU program. Graviton4 (TSMC N5, ARM Neoverse V2 core architecture) delivers approximately 30% better performance per watt than the comparable Graviton3 generation and is the default CPU for a growing share of AWS EC2 general-purpose compute. Google's Axion (also ARM Neoverse V2, TSMC N5) is deployed in Google Cloud and Google's internal infrastructure. Microsoft's Cobalt 100 (ARM Neoverse N2) entered Azure deployments in 2024 for general-purpose cloud workloads. Ampere Computing's AmpereOne (up to 192 ARM Neoverse N-class cores, TSMC N5) targets cloud-native workloads at external cloud customers who want ARM efficiency without building their own design program.

The supply chain implication is a structural demand shift at TSMC: ARM server CPU programs are consuming TSMC N5 wafer starts that would previously have gone to Intel-manufactured Xeon (Intel internal fabs) or to AMD EPYC (TSMC N5/N4, which is not new demand). The ARM server CPU share of TSMC N5 is growing every year as Graviton, Axion, and Cobalt programs scale. This adds to - not substitutes for - NVIDIA AI GPU demand at N5, tightening the overall N5 allocation picture further.

Server CPU Landscape

CPU family Supplier Architecture Foundry / node Core count (max) Primary deployment
EPYC Turin (9005) AMD Zen 5 / Zen 5c hybrid; chiplet (up to 12 compute dies + I/O die) TSMC N4P (compute dies); TSMC N6 (I/O die) 192 cores (Zen 5c dense variant) Hyperscale cloud (AWS, Google, Meta, Oracle); HPC clusters; replacing Genoa generation
EPYC Genoa (9004) AMD Zen 4; chiplet (up to 12 compute dies + I/O die); PCIe 5.0, DDR5 TSMC N5 (compute dies); TSMC N6 (I/O die) 96 cores Broad enterprise and cloud deployment; installed base largest of recent AMD EPYC generations
Xeon Granite Rapids (6) Intel P-core only (no E-core); Intel 3 process; single or dual-tile; PCIe 5.0, DDR5, CXL 2.0 Intel 3 (Intel internal fab) 128 P-cores Enterprise workloads requiring Intel platform compatibility; HPC and technical compute where Intel ecosystem matters; Intel datacenter customer base defending against AMD share gains
Xeon Clearwater Forest (next) Intel Intel 18A process; chiplet architecture; Intel's first server CPU on gate-all-around transistors Intel 18A (Intel internal - Intel Foundry Services) TBD - Intel 18A targets density improvement over Intel 3 2026+ deployment; Intel's high-stakes process recovery play; Intel 18A yield maturity is the critical supply chain variable
Graviton4 AWS (custom ARM) ARM Neoverse V2; 96 cores; 50% more cores than Graviton3; DDR5, PCIe 5.0 TSMC N5 96 cores AWS EC2 general-purpose (R8g, M8g, C8g instance families); AWS internal infrastructure
Google Axion Google (custom ARM) ARM Neoverse V2; 192 cores; optimized for Google Cloud workload mix TSMC N5 192 cores Google Cloud C4A instances; Google internal compute; YouTube, Search, Maps infrastructure
Microsoft Cobalt 100 Microsoft (custom ARM) ARM Neoverse N2; 128 cores; optimized for cloud-native and containerized workloads TSMC N5 128 cores Azure virtual machines (Cobalt 100 VM series); Microsoft Teams, Office 365 backend infrastructure
AmpereOne Ampere Computing ARM Neoverse N-class derivative; single-threaded per core (no SMT); cloud-native optimized TSMC N5 192 cores Oracle Cloud Infrastructure (OCI A1 instances); Alibaba Cloud; cloud-native customers wanting ARM without custom silicon program
Grace CPU Superchip NVIDIA ARM Neoverse V2; 72 cores; NVLink-C2C connected to Hopper or Blackwell GPU in Grace Hopper / Grace Blackwell Superchip TSMC N4 72 cores (paired with GPU in GH200 / GB200 configurations) AI training cluster host CPU (GH200, GB200 NVL72 systems); HPC supercomputers (Alps at CSCS, Eos at NVIDIA); tightly coupled CPU-GPU compute

CXL Memory — The Emerging Supply Chain

Compute Express Link (CXL) is a cache-coherent interconnect standard built on top of PCIe 5.0 physical layer that enables direct memory sharing and pooling between CPUs, GPUs, and memory expansion devices. CXL memory expansion modules - essentially DRAM attached to a server over CXL rather than via the CPU's native DRAM channels - allow datacenter operators to disaggregate compute and memory capacity, reducing stranded memory in lightly loaded servers and enabling memory pooling across multiple compute nodes. The practical effect is a new semiconductor supply chain: CXL memory module controllers, CXL switches, and CXL-attached DRAM that is distinct from both conventional server DRAM and HBM.

Samsung and SK Hynix have both announced CXL memory module products targeting datacenter deployments. Samsung's CMM-D (CXL Memory Module - DRAM) products are in early commercial deployment at cloud customers. The supply chain for CXL is still forming - CXL controller ASICs from Marvell, Montage Technology, and others are in qualification; CXL switch silicon from Microchip and Synopsys is at early commercial stage; and the PCIe 5.0 retimer ICs required for CXL signal integrity are a growing sub-category served by Parade Technologies, Montage, and Astera Labs. CXL represents a new demand vector for DRAM that is distinct from HBM (which serves accelerators) and conventional DRAM (which serves server CPU channels) - and its supply chain maturity lags approximately 2-3 years behind the standards work that preceded it.

HPC and National Supercomputer Programs

High-performance computing supercomputers represent the extreme end of datacenter semiconductor density and are the proving ground for technologies that later migrate to commercial cloud infrastructure. The US Department of Energy's Frontier supercomputer at Oak Ridge National Laboratory (peak 1.2 exaFLOPS, AMD EPYC CPUs and AMD Instinct MI250X GPUs) was the first exascale system; Aurora at Argonne National Laboratory (Intel Xeon Sapphire Rapids CPUs and Intel Ponte Vecchio GPU accelerators) is the second. Europe's Jupiter supercomputer at Forschungszentrum Jülich uses NVIDIA Grace Hopper Superchip. Japan's Fugaku (ARM A64FX processor developed by Fujitsu) demonstrated that custom ARM HPC processors can outperform x86 at scale.

The semiconductor supply chain significance of national HPC programs is threefold. First, they are the primary validation environment for new processor generations - Intel Ponte Vecchio, AMD Instinct MI250X, and NVIDIA Grace Hopper Superchip all entered volume production validated through HPC program deployments before broad commercial availability. Second, they create concentrated single-system demand for very large quantities of specific chips - a single exascale system may consume tens of thousands of accelerator cards - which can meaningfully affect supply availability for commercial customers during the procurement window. Third, sovereign HPC programs in Europe, Japan, China, and India are creating demand for domestically designed or domestically manufactured silicon that is reshaping foundry and design ecosystem investment patterns.

Datacenter Power Architecture — The 48V Transition

The AI-driven increase in per-rack power density is forcing a structural change in datacenter power delivery architecture. Conventional datacenters operate on 12V server bus power, which was adequate when rack power densities were 5-10kW. AI GPU racks - particularly NVIDIA GB200 NVL72 configurations drawing 120kW per rack - cannot be efficiently powered at 12V over the cable lengths and bus widths practical in a server rack. The industry is transitioning to 48V rack power architecture, where power is distributed at 48V to the rack and converted to lower voltages at the server or board level by GaN-based point-of-load converters.

The Open Compute Project (OCP) 48V rack power standard is the primary industry alignment mechanism for this transition. Meta, Google, Microsoft, and AWS are all OCP members and have adopted or committed to 48V rack power in new datacenter builds. The semiconductor supply chain consequence is direct: GaN power ICs (Navitas NV6128, TI LMG3522, Infineon CoolGaN) that perform the 48V-to-point-of-load conversion are the enabling device for AI-density rack power. Digital PWM controller ICs from Renesas, Monolithic Power Systems, and TI control the power conversion sequence. Vicor's 48V factorized power architecture modules are an alternative approach using proprietary power ICs. The combined GaN PSU and VRM IC demand from datacenter power architecture transition is a growing and underappreciated supply chain requirement that intersects with GaN demand from EV onboard chargers and industrial power conversion.

Supply Chain Bottlenecks and Risk Factors (2026-2030)

Bottleneck Device category Risk character Severity Resolution horizon
TSMC N5 allocation pressure ARM server CPUs (Graviton4, Axion, Cobalt, AmpereOne); AMD EPYC; NVIDIA Grace CPU ARM server CPU programs at hyperscalers growing share of TSMC N5 simultaneously with AI GPU demand growth; N5 allocation is zero-sum - ARM CPU wafer starts compete against NVIDIA GPU, AMD GPU, and mobile SoC starts for the same capacity High TSMC Arizona N4/N2 adding capacity 2025-2028; ARM server CPUs may migrate to N3/N2 nodes as those mature, freeing some N5 capacity for others; structural relief 2027-2028
Intel 18A process yield risk Intel Xeon Clearwater Forest; Intel Foundry Services customer programs Intel 18A is Intel's most critical process node in a decade - gate-all-around (GAA) transistors, PowerVia backside power delivery; if yield does not mature to commercial viability, Intel loses its internal foundry competitiveness and Xeon supply falls further behind AMD EPYC; Intel Foundry Services (IFS) external customer pipeline depends on 18A credibility High (Intel-specific) 18A risk has a binary character - either it matures (2026 target) or Intel falls back to TSMC outsourcing for Xeon as well as discrete GPU; resolution timeline is 2026-2027
800G / 1.6T optical transceiver supply Coherent optical transceivers for datacenter and AI cluster interconnect 800G transceiver supply tight relative to AI cluster buildout demand; 1.6T entering production 2025-2026 with new qualification cycle; III-V laser component supply (InP, GaAs laser diodes) is a specialty sub-supply chain with limited supplier count; co-packaged optics (CPO) transition requires new integrated photonic-electronic supply chain not yet at scale Medium-High 800G supply improving through 2025-2026; 1.6T volume 2026+; CPO requires 3-5 year supply chain development; laser component supply expansion 2-3 year lead time
GaN PSU and VRM supply GaN power ICs for 48V server PSU; VRM controllers for AI rack power delivery AI cluster power density (120kW per rack for GB200) driving GaN PSU and VRM demand far above historical server power levels; GaN supply from Navitas, Infineon, TI also serves EV OBC and industrial markets simultaneously; 48V transition creating new demand for point-of-load GaN converters across entire datacenter estate Medium-High GaN supply expanding with AI demand signal; Navitas capacity expansion at TSMC; Infineon CoolGaN production increasing; 12-18 month demand-supply lag; full relief 2026-2027 if investment decisions confirmed in 2025
DDR5 DRAM pricing cyclicality DDR5 RDIMMs for server memory DRAM pricing follows boom-bust cycles driven by capacity additions lagging demand signals; DDR4-to-DDR5 transition creates dual-inventory management complexity; AI server DDR5 demand (host memory per GPU node) growing independently of general server DDR5 and is less price-sensitive, potentially distorting standard pricing signals Medium (cyclical) Cyclical by nature - no structural resolution; three-supplier market (Samsung, SK Hynix, Micron) moderates extremes vs. single-supplier; AI server DDR5 demand providing pricing floor in down cycles
CXL ecosystem immaturity CXL controller ASICs, CXL switches, PCIe 5.0 retimer ICs CXL supply chain is early-stage - controller ASICs, switches, and retimers from multiple suppliers in simultaneous qualification; software ecosystem for CXL memory pooling not mature; hyperscaler CXL deployment expanding but at early adoption stage; supply chain fragmentation across multiple new entrants creates qualification risk Low-Medium CXL supply chain maturation 2026-2028 as hyperscaler deployments scale and qualifying suppliers consolidate; not a near-term constraint on datacenter deployment but a medium-term supply chain development requirement

Key Datacenter / HPC Semiconductor Suppliers

Company Headquarters Primary datacenter semiconductor categories Market position
Intel Santa Clara, California, US Xeon server CPU (Sapphire Rapids, Emerald Rapids, Granite Rapids, Clearwater Forest); Intel Gaudi AI accelerators; Optane persistent memory (discontinued); PCIe switch silicon; Ethernet (E810 NIC); Intel Foundry Services (IFS) Losing server CPU share to AMD EPYC and ARM alternatives; Gaudi AI accelerator gaining limited traction vs NVIDIA; Intel 18A process maturity is the critical 2026 inflection point; Intel Foundry Services is an emerging external foundry business
AMD Santa Clara, California, US EPYC server CPU (Genoa, Bergamo, Turin); Instinct GPU accelerators (MI300X, MI350X); Versal ACAP and Alveo FPGA for datacenter acceleration; ROCm open GPU compute platform Gaining server CPU share at hyperscalers and HPC; EPYC Turin (192-core Zen 5c) is the highest core-count x86 server CPU; MI300X gaining AI inference deployments at Microsoft Azure and Oracle; chiplet architecture is a manufacturing and design advantage
Broadcom San Jose, California, US Tomahawk 5 and Jericho3-AI Ethernet switch ASICs (dominant datacenter fabric); Stingray DPU; custom ASIC design services for hyperscalers (Google TPU, Meta MTIA co-design); PCIe switch ICs; SAS/SATA storage controllers Essentially every hyperscaler Ethernet fabric runs Broadcom Tomahawk ASICs; Jericho3-AI positioned for AI cluster Ultra Ethernet transition; custom ASIC revenue growing from hyperscaler silicon programs; broadest datacenter semiconductor portfolio outside CPU
Marvell Technology Santa Clara, California, US OCTEON 10 DPU; Teralynx 10 and Prestera Ethernet switch ASICs; 88SS9000 enterprise NVMe controller; optical PAM4 DSP for 800G/1.6T transceivers; custom ASIC co-development for hyperscalers; PCIe retimer ICs Second-tier Ethernet switch ASIC behind Broadcom; OCTEON DPU competes with NVIDIA BlueField; optical DSP for transceivers is a significant AI cluster dependency; custom ASIC revenue from hyperscaler programs growing and undisclosed by customer
Samsung Semiconductor Suwon, South Korea DDR5 server DRAM (largest share); CXL memory modules (CMM-D); HBM3e (qualification issues in 2024-2025, recovering); enterprise NVMe SSD (PM9A3, PM1743); NAND for enterprise SSD; Samsung Foundry SF4 (AMD dual-source) Dominant server DRAM supplier; CXL memory module market leadership; HBM3e recovery from qualification delays critical for AI cluster supply; Samsung Foundry targeting AMD and Qualcomm as anchor customers for SF3/SF2 nodes
Micron Technology Boise, Idaho, US DDR5 server DRAM; HBM3e (qualified at NVIDIA 2024, ramping); enterprise NVMe SSD (9400 series); LPDDR5X for low-power server variants; CXL memory module development Only US-headquartered DRAM and NAND supplier; HBM3e qualification at NVIDIA provides AI supply chain revenue and strategic importance; Manassas VA DRAM fab serves defense and aerospace ECC memory market
Arm Holdings Cambridge, United Kingdom Neoverse V2 and N2 CPU IP cores (licensed to AWS, Google, Microsoft, Ampere, NVIDIA, Qualcomm); CSS (Compute Subsystem) reference designs; Mali and Immortalis GPU IP; CoreLink interconnect IP The foundational IP licensor for all ARM server CPU programs; ARM's royalty revenue grows with every Graviton, Axion, Cobalt, and AmpereOne deployment; ARM server CPU market share growth directly translates to ARM Holdings revenue growth
Ampere Computing Santa Clara, California, US AmpereOne ARM server CPU (up to 192 cores); AmpereOne-X with integrated AI inference accelerator; cloud-native optimized server CPU with no SMT The external ARM server CPU option for cloud customers not building their own; Oracle Cloud Infrastructure and Alibaba Cloud are anchor customers; competes against hyperscaler in-house ARM CPUs at their own cloud but wins external cloud operators
Coherent Corp Saxonburg, Pennsylvania, US Coherent optical transceiver DSP ASICs; 400G, 800G, and 1.6T pluggable optical modules; silicon photonics; compound semiconductor laser and modulator components; indium phosphide (InP) wafer production One of two dominant coherent optical transceiver suppliers (with Lumentum/Viavi); AI cluster 800G transceiver demand is Coherent's largest revenue growth driver; InP compound semiconductor vertical integration is a supply chain moat

Cross-Sector Convergence

Datacenter semiconductors intersect three significant cross-sector supply chain dynamics. First, the TSMC N5 allocation competition: ARM server CPUs at hyperscalers (Graviton4, Axion, Cobalt, AmpereOne) compete for TSMC N5 wafer starts against AI GPU programs, mobile SoCs, automotive ADAS compute, and AMD's own EPYC client programs. As ARM server CPU deployments scale from 30-40% of new hyperscaler CPU installations today toward potential majority share by 2028-2030, the TSMC N5 demand from this category alone will be structurally significant even before accounting for AI and mobile.

Second, the GaN power delivery convergence: the datacenter transition to 48V rack power architecture drives GaN PSU IC demand from the same supply base serving EV onboard chargers, industrial power conversion, and robot joint motor drives. GaN is not a commodity - it is a specialty semiconductor manufactured at TSMC and TI internal fabs with limited capacity expansion pace. The simultaneous surge in datacenter GaN demand (48V AI rack PSU) and EV GaN demand (800V OBC) and robotics GaN demand (joint drive ICs) is not coordinated and is not reflected in any single supplier's disclosed capacity expansion plan. The combined demand trajectory makes GaN capacity expansion decisions in 2025-2026 critically important across multiple sectors simultaneously.

Third, the optical interconnect compound semiconductor supply: coherent optical transceivers for datacenter interconnect use indium phosphide (InP) laser diodes and electro-absorption modulators manufactured at III-V compound semiconductor fabs - a separate and smaller supply chain than silicon. Coherent Corp, Lumentum, and II-VI/Coherent are the primary InP component suppliers. As datacenter bandwidth scaling accelerates with AI cluster buildout (800G to 1.6T to co-packaged optics), InP laser component supply becomes a binding constraint upstream of the transceiver DSP ASIC supply chain. This is the optical equivalent of the SiC substrate constraint in power semiconductors - a physical material supply limit that cannot be resolved by adding CMOS foundry capacity.

Related Coverage: CPUs | Memory & Storage | HBM | FPGAs | ASICs | CoWoS Packaging | Bottleneck Atlas | AI & ML | 5G/6G & Wireless | SiC & GaN Power Modules

Cross-Network: ElectronsX Demand Side

Datacenter semiconductor demand intersects EX coverage at the data center power infrastructure layer - the grid connection, UPS systems, cooling infrastructure, and on-site generation that enable hyperscale compute at the power densities AI clusters require.

EX: Supply Chain Convergence Map | EX: BESS Supply Chain | EX: Electrification Bottleneck Atlas

Key Questions — Datacenter / HPC Semiconductors

Is ARM actually displacing x86 in the datacenter? At hyperscalers, yes - and the displacement is structural rather than cyclical. AWS, Google, and Microsoft have each committed capital and engineering resources to multi-year ARM CPU programs that cannot be easily reversed. The economics are compelling: ARM server CPUs deliver 30-40% better performance per watt at cloud-native workloads, and hyperscalers operate at a scale where electricity cost differences translate directly to hundreds of millions of dollars in annual savings. For enterprise datacenters with large x86 software portfolios and ISV certification requirements, x86 will remain dominant through 2030 - ARM's datacenter disruption is concentrated at the hyperscaler layer where workloads are containerized and software compatibility is less constraining.

What is the significance of Intel 18A for the server CPU market? Intel 18A is Intel's attempt to recover process leadership using gate-all-around (GAA) transistors and PowerVia backside power delivery on the same chip - two major architectural innovations simultaneously. If Intel 18A matures to commercial yield by 2026, Xeon Clearwater Forest could close the performance-per-watt gap with AMD EPYC Turin and TSMC-manufactured ARM server CPUs. If Intel 18A yield does not mature, Intel will need to outsource Xeon production to TSMC, which adds TSMC N3/N2 demand and further tightens leading-edge foundry allocation - while also being a strategic defeat for Intel's foundry ambitions. The 18A outcome is the single most consequential binary event in server CPU supply chains through 2027.

How does the datacenter sector relate to the AI sector on this site? The AI sector page covers the accelerator supply chain: GPU dies (TSMC N3/N5), HBM memory stacks (SK Hynix, Samsung, Micron), CoWoS advanced packaging (TSMC), and custom hyperscaler ASICs. This page covers the surrounding infrastructure semiconductor stack: the host server CPUs that manage GPU nodes, the DDR5 DRAM that buffers data for GPU processing, the Ethernet switch ASICs that interconnect GPU servers, the DPUs that offload networking from host CPUs, and the GaN power delivery ICs that feed the whole system. These are distinct supply chains with distinct constraints and distinct lead times, but they must scale together - a datacenter operator cannot deploy GPU capacity without the surrounding infrastructure silicon.

What is co-packaged optics and why does it matter for supply chains? Co-packaged optics (CPO) integrates optical transceiver functionality directly into the switch ASIC package rather than using pluggable optical modules connected via a faceplate port. The advantage is bandwidth density and power efficiency - CPO eliminates the SerDes electrical interface between switch ASIC and pluggable module, which becomes the power and bandwidth bottleneck at 1.6T and beyond. The supply chain implication is significant: CPO requires the co-design and co-manufacturing of silicon photonic integrated circuits (PICs) with switch ASICs in the same package, creating a new class of semiconductor product that bridges CMOS foundry supply chains and III-V compound semiconductor supply chains. Neither Broadcom, Marvell, nor the optical transceiver suppliers have fully mature CPO supply chains - which is why the transition from pluggable to CPO will take until 2028-2030 even as the bandwidth requirements are pressing against pluggable limits today.

Related Coverage

Sectors Hub | CPUs | Memory & Storage | HBM | FPGAs | ASICs | GPU & CPU Boards | Memory Modules | CoWoS Packaging | Bottleneck Atlas | AI & ML | 5G/6G & Wireless | PC | SiC & GaN Power Modules | Fab Power