Hardware for HPC and AI 2025-2035: Technologies, Markets, Forecasts

High-Performance Computing, Exascale Supercomputers, AI Chips, CPUs, GPUs and Accelerators, Memory, Storage, Advanced Semiconductor Packaging, Networking, Interconnects and Thermal Management.

Show All Description Contents, Table & Figures List Pricing Related Content
This report examines the high-performance computing (HPC) and AI hardware markets. It provides an overview of the overall HPC market in the exascale era, including analysis of major supercomputers like El Capitan. Detailed overview of technologies including GPUs and accelerators, CPUs, memory, advanced semiconductor packaging, interconnects, and thermal management is provided, alongside over 50 profiles of HPC industry players, many from primary interviews. Highly granular forecasts from 2025 to 2035 in volume and revenue for the overall HPC and AI hardware market and 6 key classes of hardware provide a picture of the evolving opportunities in this market, with a US$581 billion market value forecast by 2035.
 
HPC market, Supercomputing industry, HPC forecast, High Performance Computing technology, AI computing trends
 
HPC systems, including supercomputers, outclass all other classes of computing in terms of calculation speed by parallelizing processing over many processors. HPC has long been an integral tool across critical industries, from facilitating engineering modelling to predicting the weather. As both the use-cases of AI and machine learning have grown alongside the scale of their underlying models, the demand for HPC systems has only intensified, alongside the demands on their processing power and interconnects between computing nodes.
 
The largest HPC systems in the world, which can perform over 1 quintillion floating-point operations per second, are the preserve of government-operated national labs of global superpowers. However, private corporations are not far behind, with cloud providers like Microsoft Azure possessing supercomputers in the upper echelons of performance. Growth here is only expected to accelerate as the deployment of AI widens.
 
IDTechEx's report examines the key hardware challenges in HPC, which revolve around processors, memory and storage, advanced packaging, interconnects, and thermal management. GPUs and other AI accelerators grab all the headlines in the processor space, with important trends including the adoption of cores for tensor processing within GPUs and hints of AMD challenging NVIDIA. However, substantial progress has also taken place in the CPU market, with core counts approaching 200 in chips like AMD's Turin processors and the deployment of integrated heterogeneous systems - GPUs and CPUs integrated into the same package - taking place in the world's most powerful supercomputers, including the USA's El Capitan and China's Tianhe-3.
 
In the memory space, the adoption of HBM (High Bandwidth Memory) has been crucial to the growing capabilities of accelerators, with approximately 95% of accelerators in HPC now employing the technology, but research into next-generation memory technologies such as selector-only memory (SOM), phase change memory (PCRAM) and magnetoresistive RAM (MRAM) is being driven by the high energy consumption of today's memory choices. The quest for high-capacity storage that meets the increasing performance requirements of HPC workloads is seeing the emergence of high-density QLC SSDs at a lower cost-point than typical TLC SSDs available in the market.
 
Advanced packaging is a key enabler of all these chip technologies, with the 3D stacking of memory on top of logic being the best approach to achieve ultra-high bandwidths by shrinking on-device interconnect distances. Advanced packaging techniques are also highly important for the roadmaps of the interconnects between HPC nodes, as future iterations of low-latency networks like NVIDIA's InfiniBand and HPE's Slingshot move towards co-packaged optics to further reduce the distances data must travel at the chip scale and speed up throughput overall.
 
Of course, these advances evolve large amounts of heat, so thermal management performance is key. Despite immersion cooling's strong performance in operational expense and performance, cold plate cooling is expected to remain the technology of choice in most cases as HPC builders seek to reduce the massive capital expenses that come with building these enormous interconnected machines.
 
HPC market, Supercomputing industry, HPC forecast, High Performance Computing technology, AI computing trends
 
For those needing to understand the HPC market, including the above trends and further challenges like the development of storage and the geopolitical dynamics that have affected the growth of this technology, IDTechEx's report is a must. At the dawn of the exascale era and in the most fruitful period for AI development yet, the demand for supercomputing is set for rapid expansion. As such, Hardware for HPC and AI 2025-2035 includes 10-year forecasts not only for the overall HPC hardware market, but for key hardware technologies including CPUs, GPUs, memory including HBM, storage and server boards. The reader will leave equipped with a wide-ranging, in-depth picture of the present and future of this exciting industry.
 
Key aspects
This report from IDTechEx covers the following key contents:
 
High-performance computing architectures and supercomputers:
A) Defining key architectures for high-performance computing and key industry trends including the rise of exascale computers.
B) Exploring the current state of HPC development, including market drivers and trends, case studies, major OEMs, the influence of international trade, and the HPC value chain.
 
Technology developments of hardware for HPC & AI:
A) Exploring current developments of processing units, such as CPUs, GPUs, and other application-specific hardware for high-performance computing and their development based on the increasing need for AI within these systems.
B) Investigating technologies for memory (RAM) for high-performance computing and AI, such as high bandwidth memory, DDR memory, and emerging technologies.
C) Exploring storage technologies in high-performance computing. Focusing on advanced storage, like NAND flash memory and its use cases in storage servers. Includes analysis of storage technologies for various workloads.
D) Investigating advanced semiconductor technologies used in high-performance computing systems, such as the move into chiplet integration using both 2.5D and 3D packaging technologies.
E) Breakdown of networking and interconnects in high-performance computing, such as copper and optical-based solutions like pluggable and co-packaged optics.
F) Review of types of cooling systems in data center systems used for high-performance computing and technology trends based on the growing need for high-power systems.
G) All sections cover comparative analysis, key trends, case studies, technology benchmarks, and developments for key players.
 
Market forecasts & analysis:
A) 10-year granular market forecasts separated by key hardware used for HPC systems.
B) Assessment of key technological and commercial trends for different hardware in HPC systems.
Report MetricsDetails
CAGRHPC hardware market to reach US$581 bn by 2035 at a CAGR of 13.6%.
Forecast Period2025 - 2035
Forecast Units- Annual revenue (US$ bn) - Yearly unit sales (millions)
Regions CoveredWorldwide
Segments Covered- Server units - CPUs - GPUs - High Bandwidth Memory (HBM) - Memory and Storage - Overall HPC hardware - Thermal management methods
Analyst access from IDTechEx
All report purchases include up to 30 minutes telephone time with an expert analyst who will help you link key findings in the report to the business issues you're addressing. This needs to be used within three months of purchasing the report.
Further information
If you have any questions about this report, please do not hesitate to contact our report team at research@IDTechEx.com or call one of our sales managers:

AMERICAS (USA): +1 617 577 7890
ASIA (Japan): +81 3 3216 7209
EUROPE (UK) +44 1223 812300
1.EXECUTIVE SUMMARY
1.1.What is high performance computing (HPC)?
1.2.HPC architectures
1.3.On-premises, cloud, and hybrid HPC infrastructures: shift towards cloud & hybrid
1.4.Development to exascale supercomputers and beyond
1.5.Distribution of TOP500 supercomputers (June 2024)
1.6.Exascale supercomputers around the world
1.7.The HPC manufacturer landscape
1.8.Supercomputing in China and the US-China trade war (I): no more contribution to the TOP500 list
1.9.Supercomputing in China and the US-China trade war (II): onshoring chip supply
1.10.State of the art exascale supercomputers: Aurora, Frontier and El Capitan
1.11.El Capitan: the top US HPC in November 2024
1.12.Overview of players in HPC
1.13.CPUs for HPC and AI workloads introduction
1.14.CPU introduction
1.15.Core Architecture in a CPU
1.16.Key CPU Requirements for HPC and AI Workloads (1)
1.17.Key CPU Requirements for HPC and AI Workloads (2)
1.18.HPC CPU Market Trends by Company in the TOP500 Supercomputers
1.19.HPC Market Trends by Company in the TOP500 Supercomputers
1.20.CPU and GPU Architecture Comparison
1.21.High Performance GPU Architecture Breakdown
1.22.Highlights from Key GPU Players
1.23.Trends in Data Center GPUs
1.24.2.5D and 3D semiconductor packaging technologies key to HPC and AI
1.25.Key metrics for advanced semiconductor packaging performance
1.26.Overview of trends in HPC chip integration
1.27.Advanced semiconductor packaging path for HPC
1.28.DDR memory dominates CPUs whereas HBM is key to GPU performance
1.29.Forecast: Yearly Unit Sales and Market Size of High Bandwidth Memory (HBM)
1.30.Forecasts: Memory and Storage for Servers
1.31.Memory bottlenecks are driving segmentation of memory hierarchy
1.32.Data movement through storage tiers in clusters for AI workloads
1.33.TRL of storage technologies for HPC Datacenter and Enterprise
1.34.TDP Trend: Historic Data and Forecast Data - CPU
1.35.Cooling Methods Overview for High Performance Computers
1.36.Different Cooling on The Chip Level
1.37.Yearly Revenue Forecast By Cooling Method: 2022-2035
1.38.Forecasts: HPC Hardware Market Size
2.INTRODUCTION
2.1.HPC - overview
2.1.1.The growing demand for data computing power
2.1.2.Integrated circuits explained
2.1.3.The need for specialized chips
2.1.4.Fundamentals of abundant data computing system
2.1.5.Memory bandwidth deficit
2.1.6.Four key area of growth for abundance data computing system
2.1.7.Key parameters for transistor device scaling
2.1.8.Evolution of transistor device architectures
2.1.9.Scaling technology roadmap overview
2.1.10.Process nodes
2.1.11.Semiconductor foundries and their roadmap
2.1.12.Roadmap for advanced nodes
2.1.13.The rise of advanced semiconductor packaging and its challenges
2.2.Data centers
2.2.1.Introduction to Data Centers
2.2.2.Data Center Equipment - Top Level Overview
2.2.3.Data Center Server Rack and Server Structure
2.2.4.Data Center Switch Topology - Three Layer and Spine-Leaf Architecture
2.2.5.K-ary Fat Tree Topology
2.2.6.TDP Increases Over Time - GPU
2.2.7.Server Boards TDP Increases - Moore's Law
2.3.What is HPC?
2.3.1.High performance computing (HPC)
2.3.2.Types of High Performance Computing (HPC) Architecture
2.3.3.HPC architectures
2.3.4.Supercomputers and mainframes for HPC
2.3.5.On-premises, cloud and hybrid HPC infrastructures: shift towards cloud & hybrid
2.3.6.Fundamentals of abundance data computing system
2.3.7.Development to exascale supercomputers and beyond
2.3.8.Distribution of TOP500 supercomputers (June 2024)
2.3.9.Limitations of the TOP500 list
2.4.Chip technologies for HPC
2.4.1.The rise and the challenges of LLM
2.4.2.What is an AI chip?
2.4.3.AI acceleration
2.4.4.Why AI acceleration is needed
2.4.5.The AI chip landscape - overview
2.4.6.State-of-the-art high-end AI and HPC chips
2.4.7.Evolution of AI compute system architecture
2.4.8.NVIDIA and AMD solutions for next-gen AI
2.4.9.The next step forward to improve system bandwidth
2.5.Benchmarking HPC
2.5.1.HPL
2.5.2.HPC benchmarking methods overview
2.5.3.Comparison of Energy Efficiency and Performance of Supercomputers
2.5.4.HPL-AI
2.5.5.MLPerf by MLCommons
2.5.6.MLPerf - HPC
2.5.7.MLPerf Inference- Datacenter
2.5.8.Adoption of Accelerators/Co-Processors in Supercomputers is Growing but Far from Universal
2.5.9.Growth in Accelerator Adoption within the TOP500 is Highest in the Vendor Segment
2.6.HPC Installation Case Studies
2.6.1.Exascale supercomputers around the world
2.6.2.The HPC manufacturer landscape
2.6.3.Lenovo's global reach is unique
2.6.4.The MareNostrum 5's dual-purpose architecture
2.6.5.Fujitsu pioneered ARM architectures in HPC
2.6.6.The Fugaku HPC at Riken
2.6.7.Supercomputing in China and the US-China trade war (I): no more contribution to the TOP500 list
2.6.8.Supercomputing in China and the US-China trade war (II): onshoring chip supply
2.6.9.Tianhe-3/Xingyi: China's most powerful HPC
2.6.10.HPE is a major force in the global supercomputing market
2.6.11.Overview of the node architecture on the Frontier supercomputer
2.6.12.State of the art exascale supercomputers: Aurora, Frontier and El Capitan
2.6.13.El Capitan: the top US HPC in November 2024
2.6.14.Overview of players in HPC
3.FORECASTS
3.1.Forecast Methodology
3.2.Forecasts: HPC Server Unit Sales
3.3.Forecast: Unit Sales of CPUs and GPUs for HPC
3.4.Forecast: Market Size of CPUs and GPUs for HPC
3.5.Forecast: Yearly Unit Sales and Market Size of High Bandwidth Memory (HBM)
3.6.Forecasts: Memory and Storage for Servers
3.7.Forecasts: HPC Hardware Market Size
3.8.Yearly Revenue Forecast By Cooling Method: 2022-2035
4.CPUS FOR HPC AND AI
4.1.Overview
4.1.1.CPUs for HPC and AI Workloads Introduction
4.1.2.CPU Introduction
4.1.3.Players: CPUs for Server Processors
4.1.4.Core Architecture in a CPU
4.1.5.The Instruction Cycle
4.2.Key Requirements for CPUs in HPC and AI Workloads and Technology Landscape
4.2.1.Key CPU Requirements for HPC and AI Workloads (1)
4.2.2.Key CPU Requirements for HPC and AI Workloads (2)
4.2.3.Intel CPUs found in Server Processors: Current Landscape
4.2.4.Intel CPUs found in Server Processors: Technology Overview
4.2.5.Intel CPUs get Built-In AI Accelerators with Intel Advanced Matrix Extensions
4.2.6.AMD CPUs found in Server Processors: Current Landscape
4.2.7.AMD CPUs found in Server Processors: Technology Overview
4.2.8.AMD uses 3D V-Cache Technology for 1152MB L3 Cache
4.2.9.AVX-512 Vector Extensions for x86-64 Instruction Set
4.2.10.IBM Power CPUs found in Server Processors: Current Landscape
4.2.11.Arm Holdings Licences Core Designs with its RISC-Based Instruction Sets
4.2.12.Arm based A64FX Fujitsu Processors for Supercomputers
4.2.13.Chinese Sunway Processors Perform well despite Trade Restrictions
4.3.CPU Market Trends
4.3.1.Intel Xeon and AMD EPYC Uptake in TOP500 Supercomputers
4.3.2.HPC Market Trends by Company in the TOP500 Supercomputers
4.3.3.Processor Comparison: TDP, Price and Core Count (1)
4.3.4.Processor Comparison: TDP, Price and Core Count (2)
4.3.5.RISC Adoption Consideration and Specialised GPUs
5.GPUS AND OTHER AI ACCELERATORS
5.1.GPUs Background
5.1.1.Importance of GPU Software Stacks
5.1.2.High Performance GPU Architecture Breakdown
5.1.3.CPU and GPU Architecture Comparison
5.1.4.Integrated Heterogeneous Systems
5.1.5.Highlights from Key GPU Players
5.2.GPU Case Studies
5.2.1.NVIDIA: NVIDIA CUDA and Tensor Cores
5.2.2.NVIDIA: Blackwell GPU
5.2.3.NVIDIA Rack-Scale Solutions for Datacenters
5.2.4.AMD: CDNA 3 Architecture and Compute Units for GPU Compute
5.2.5.AMD: AMD Instinct MI300A Data Centre APUs are Heterogeneous Systems
5.2.6.Intel: Intel GPU Max and the Xe-HPC Architecture
5.2.7.Intel: Falcon Shores Integrates HPC and AI GPU with AI Accelerator Tech
5.2.8.Precision Performance for HPC and AI Workloads
5.2.9.Benchmark
5.3.GPU Market Insights
5.3.1.HPC Accelerator Market Trends by Company in the TOP500 Supercomputers
5.3.2.HPC Accelerator Market Trends by Company in the TOP500 Supercomputers
5.3.3.Datacenter GPUs Use Cases and Business Models
5.4.Trends and Solutions In GPU Technology
5.4.1.Trends in Data Center GPUs
5.4.2.Transistor Technologies Advance for High-Performance GPUs
5.4.3.Challenges in Transistor Scaling
5.4.4.Overcoming the Reticle Limit for Die Scaling
5.4.5.Case Study: Cerebras Overcomes Die Scaling using Single-Chip Wafers
5.4.6.Tachyum Builds from the Ground up to Deliver Exceptional Performance
5.4.7.Tachyum: Barriers to Entry for Startups
5.5.Application-Specific Hardware for Data Centers
5.5.1.Application-Specific Hardware Provides Solution for Slowing Moore's Law in GPUs for HPC and AI
5.5.2.AI Accelerators
5.5.3.AI chip types
5.5.4.Accelerators in servers
5.5.5.Main Players for AI Accelerators
5.6.Summary: GPUs and Accelerators
5.6.1.GPU and AI Accelerators Summary
6.MEMORY (RAM) FOR HPC AND AI
6.1.Overview
6.1.1.Hierarchy of computing memory
6.1.2.Memory bottlenecks for HPC/AI workloads and processor under-utilization
6.1.3.Types of DRAM and Comparison of HBM with DDR
6.1.4.Memory technology in TOP500 supercomputers processors and accelerators
6.1.5.HBM vs DDR for computing - market trend
6.1.6.Demand outgrows supply for HBM in 2024
6.2.High Bandwidth Memory (HBM)
6.2.1.High bandwidth memory (HBM) and comparison with other DRAM technologies
6.2.2.HBM (High Bandwidth Memory) packaging
6.2.3.HBM packaging transition to hybrid bonding
6.2.4.Benchmark of HBM performance utilizing µ bump and hybrid bonding
6.2.5.SK Hynix has started volume production of 12-layer HBM3E
6.2.6.Micron released 24GB HBM3E for NVIDIA H200 and is sampling 36GB HBM3E
6.2.7.Samsung expects production of HBM3E 36GB within 2024
6.2.8.Overview of current HBM stacking technologies by 3 main players
6.2.9.Evolution of HBM generations and transition to HBM4
6.2.10.Benchmarking of HBM technologies in the market from key players (1)
6.2.11.Benchmarking of HBM technologies in the market from key players (2)
6.2.12.Examples of CPUs and accelerators using HBM
6.2.13.Intel's CPU Max series for HPC workloads has HBM and optional DDR
6.2.14.AMD CDNA 3 APU architecture with unified HBM memory for HPC
6.2.15.Three main approaches to package HBM and GPU
6.2.16.Drawbacks of High Bandwidth Memory (HBM)
6.3.DDR Memory
6.3.1.Developments in double data rate (DDR) memory
6.3.2.DDR5 memory in AMD's 4th Gen EPYC processors for HPC workloads
6.3.3.DDR5 MRDIMM increases capacity and bandwidth for high CPU core counts
6.3.4.NVIDIA's Grace CPU uses LPDDR5X memory to lower power consumption
6.3.5.GDDR7 announced by major players targeting HPC and AI applications
6.3.6.Comparison of GDDR6 and GDDR7 modules
6.4.Memory Expansion
6.4.1.Samsung's CMM-D memory expansion for AI and datacenter server applications
6.4.2.Micron's CXL memory expansion modules for storage tiering in datacenters
6.5.Emerging memory technologies
6.5.1.Shortfalls of DRAM is driving research into next-generation memory solutions
6.5.2.Phase change memory and phase change materials for 3DXP
6.5.3.Selector only memory (SOM) for storage class memory solutions
6.5.4.Magnetic Random Access Memory (MRAM) can boost energy-efficiency
6.5.5.Resistive Random Access Memory (ReRAM)
6.6.Market and Outlook
6.6.1.DDR memory dominates CPUs whereas HBM is key to GPU performance
6.6.2.Memory bottlenecks are driving segmentation of memory hierarchy
6.6.3.Memory and Storage: market status and outlook
7.STORAGE FOR HPC
7.1.Overview
7.1.1.Flash storage is the leading storage technology for HPC and AI applications
7.1.2.HPC and AI require large-scale and high-performance data storage
7.1.3.Storage requirements varies depending on the AI workloads
7.1.4.Data movement through storage tiers in clusters for AI workloads
7.2.NAND Flash SSD Technologies for Datacenter and Enterprise
7.2.1.NAND Flash memory uses floating gates or charge traps to store data
7.2.2.Data center and enterprise SSD form factors transitioning towards EDSFF
7.2.3.SK Hynix - NAND technology development
7.2.4.SK Hynix: overcoming stacking limitations to increase capacity using 4D2.0
7.2.5.Examples of SK Hynix NAND Flash storage chips for AI and data centers
7.2.6.Solidigm (SK Hynix subsidiary) offers SSDs previously manufactured by Intel
7.2.7.Micron's 9550 SSDs are designed for AI-critical workloads with PCIe Gen5
7.2.8.Micron has a range of SSDs for applications in datacenters and AI
7.2.9.Example of SSD configurations and solutions for AI and HPC workloads
7.2.10.KIOXIA uses BiCS 3D FLASHTM Technology to increase storage density
7.2.11.KIOXIA offers a range of datacenter and enterprise SSD solutions
7.2.12.Samsung's 9th generation V-NAND technology is targeting AI applications
7.2.13.Intel's Optane SSD latency-sensitive solutions were designed for server caching
7.3.Technology Comparison and Trends
7.3.1.Increasing SSD capacity through emerging lower cost QLC NAND
7.3.2.QLC affords higher capacity at a lower cost per bit but with performance deficits
7.3.3.Step change in sequential read bandwidth with each generation of PCIe
7.3.4.Evolution of PCIe generations in the SSD market
7.3.5.SSDs performance for different datacenter and enterprise applications
7.3.6.SSDs for storage class memory bridging gap to volatile memory
7.3.7.Vertical stacking is unlikely to be the only factor in increasing storage density
7.3.8.Energy efficiency of SSD storage technologies for datacenter and enterprise
7.3.9.Examples of SSDs for AI workloads
7.4.Storage servers
7.4.1.HPE's Cray ClusterStor E1000 for HPC applications
7.4.2.Storage capacity of Cray ClusterStor E1000 is scalable according to configuration
7.4.3.NetApp offers all-flash storage systems for datacenters and enterprises
7.4.4.Examples of other all-Flash scalable storage systems
7.5.Players and Outlook
7.5.1.Examples of players in HPC storage
7.5.2.TRL of storage technologies for HPC Datacenter and Enterprise
8.ADVANCED SEMICONDUCTOR PACKAGING TO ENABLE HPC
8.1.Overview of Advanced Semiconductor Packaging Technologies
8.1.1.Progression from 1D to 3D semiconductor packaging
8.1.2.Key metrics for advanced semiconductor packaging performance
8.1.3.Four key factors of advanced semiconductor packaging
8.1.4.Overview of interconnection technique in semiconductor packaging
8.1.5.Overview of 2.5D packaging structure
8.1.6.2.5D advanced semiconductor packaging technology portfolio
8.1.7.Evolution of bumping technologies
8.1.8.3D advanced semiconductor packaging technology portfolio
8.2.How can advanced semiconductor packaging address these challenges?
8.2.1.Why traditional Moore's Law scaling cannot meet the growing needs for HPC
8.2.2.Key Factors affecting HPC datacenter performance
8.2.3.Advanced semiconductor packaging path for HPC
8.2.4.HPC chips integration trend - overview
8.2.5.HPC chips integration trend - explanation
8.2.6.Enhancing Energy Efficiency Through Hybrid Bonding Pitch Scaling
8.2.7.What is critical in packaging for AI & HPC
8.2.8.Future system-in-package architecture
8.3.Case studies
8.3.1.AMD Instinct MI300
8.3.2.AMD Instinct MI300 is being used for exascale supercomputing
8.3.3.3D chip stacking using hybrid bonding - AMD
8.3.4.Advanced semiconductor packaging for Intel latest Xeon server CPU (1)
8.3.5.Xilinx FPGA packaging
8.3.6.Future packaging trend for chiplet server CPU
8.3.7.High-end commercial chips based on advanced semiconductor packaging technology (1)
8.3.8.High-end commercial chips based on advanced semiconductor packaging technology (2)
9.NETWORKING AND INTERCONNECT IN HPC
9.1.Overview
9.1.1.Introduction: networks in HPC
9.1.2.Switches: Key components in a modern data center
9.1.3.Why does HPC require high-performance transceivers?
9.1.4.Current AI system architecture
9.1.5.L1 backside compute architecture with Cu systems
9.1.6.Number of Cu wires in current AI system Interconnects
9.1.7.Limitations in current copper systems in AI
9.1.8.What is an Optical Engine (OE)
9.1.9.Integrated Photonic Transceivers
9.1.10.Nvidia's connectivity choices: Copper vs. optical for high-bandwidth systems
9.1.11.Pluggable optics vs CPO
9.1.12.L1 backside compute architecture with optical interconnect: Co-Packaged Optics (CPO)
9.1.13.Key CPO applications: Network switch and computing optical I/O
9.1.14.Design decisions for CPO compared to Pluggables
9.1.15.Power efficiency comparison: CPO vs. Pluggable optics vs. copper interconnects
9.1.16.Latency of 60cm data transmission technology benchmark
9.1.17.Future AI architecture predicted by IDTechEx
9.2.Optical on-device interconnects
9.2.1.On-device interconnects
9.2.2.Overcoming the von Neumann bottleneck
9.2.3.Electrical Interconnects Case Study: NVIDIA Grace Hopper for AI
9.2.4.Why improve on-device interconnects?
9.2.5.Improved Interconnects for Switches
9.2.6.Case Study: Ayer Labs TeraPHY
9.2.7.Case Study: Lightmatter's 'Passage' PIC-based Interconnect
9.3.Interconnect standards in HPC
9.3.1.The role of interconnect families in HPC
9.3.2.Lanes in interconnect
9.3.3.Comparing roadmaps
9.3.4.NVIDIA and the resurgence in InfiniBand interconnects
9.3.5.InfiniBand is more dominant when considering new-build HPCs
9.3.6.Intel, Cornelis Networks and Omni-Path
9.3.7.Cornelis Networks and the new Omni-Path CN5000 standard
9.3.8.Gigabit ethernet: the open standard for HPC interconnect
9.3.9.HPE Cray's Slingshot interconnect
9.4.Summary: Networking in HPC
9.4.1.Summary: networking for HPC
10.THERMAL MANAGEMENT FOR HPC
10.1.Increasing TDP Drives More Efficient Thermal Management
10.2.Historic Data of TDP - GPU
10.3.TDP Trend: Historic Data and Forecast Data - CPU
10.4.Data Center Cooling Classification
10.5.Cooling Methods Overview
10.6.Different Cooling on Chip Level
10.7.Cooling Technology Comparison
10.8.Cold Plate/Direct to Chip Cooling - Standalone Cold Plate
10.9.Liquid Cooling Cold Plates
10.10.Single and Two-Phase Cold Plate: Challenges
10.11.Single-Phase and Two-Phase Immersion - Overview (1)
10.12.Single and Two-Phase Immersion: Challenges
10.13.Coolant Comparison
10.14.Coolant Comparison - PFAS Regulations
10.15.Coolant Distribution Units (CDU)
10.16.Heat Transfer - Thermal Interface Materials (TIMs) (1)
10.17.Heat Transfer - Thermal Interface Materials (TIMs) (2)
11.COMPANY PROFILES
11.1.Accelsius — Two-Phase Direct-to-Chip Cooling
11.2.ACCRETECH (Grinding Tool)
11.3.AEPONYX
11.4.Amazon AWS Data Center
11.5.Amkor — Advanced Semiconductor Packaging
11.6.Arieca (2024)
11.7.Arieca (2020)
11.8.ASE — Advanced Semiconductor Packaging
11.9.Asperitas Immersed Computing
11.10.Calyos: Data Center Applications
11.11.Camtek
11.12.CEA-Leti (Advanced Semiconductor Packaging)
11.13.Ciena
11.14.Coherent: Photonic Integrated Circuit-Based Transceivers
11.15.Comet Yxlon
11.16.DELO: Semiconductor Packaging
11.17.EFFECT Photonics
11.18.Engineered Fluids
11.19.EU Chips Act — SemiWise
11.20.EVG (D Hybrid Bonding Tool)
11.21.GlobalFoundries
11.22.Green Revolution Cooling (GRC)
11.23.Henkel: microTIM and data centers
11.24.JCET Group
11.25.JSR Corporation
11.26.Kioxia
11.27.Lightelligence
11.28.Lightmatter
11.29.LioniX
11.30.LIPAC
11.31.LiquidCool Solutions — Chassis-Based Immersion Cooling
11.32.LiSAT
11.33.LPKF
11.34.Lumiphase
11.35.Micron
11.36.Mitsui Mining & Smelting (Advanced Semiconductor Packaging)
11.37.Nano-Join
11.38.NanoWired
11.39.NeoFan
11.40.Neurok Thermocon Inc
11.41.Parker Lord: Dispensable Gap Fillers
11.42.Resonac (RDL Insulation Layer)
11.43.Resonac Holdings
11.44.Samsung Electronics (Memory)
11.45.SK Hynix
11.46.Solidigm
11.47.Sumitomo Chemical Co, Ltd
11.48.Taybo (Shanghai) Environmental Technology Co, Ltd
11.49.TDK
11.50.TOK
11.51.TSMC (Advanced Semiconductor Packaging)
11.52.Tyson
11.53.Vertiv Holdings - Data Center Liquid Cooling
11.54.Vitron (Through-Glass Via Manufacturing) — A LPKF Trademark
11.55.ZutaCore
 

Ordering Information

Hardware for HPC and AI 2025-2035: Technologies, Markets, Forecasts

£$¥
Electronic (1-5 users)
£5,650.00
Electronic (6-10 users)
£8,050.00
Electronic and 1 Hardcopy (1-5 users)
£6,450.00
Electronic and 1 Hardcopy (6-10 users)
£8,850.00
Electronic (1-5 users)
€6,400.00
Electronic (6-10 users)
€9,100.00
Electronic and 1 Hardcopy (1-5 users)
€7,310.00
Electronic and 1 Hardcopy (6-10 users)
€10,010.00
Electronic (1-5 users)
$7,000.00
Electronic (6-10 users)
$10,000.00
Electronic and 1 Hardcopy (1-5 users)
$7,975.00
Electronic and 1 Hardcopy (6-10 users)
$10,975.00
Electronic (1-5 users)
元50,000.00
Electronic (6-10 users)
元72,000.00
Electronic and 1 Hardcopy (1-5 users)
元58,000.00
Electronic and 1 Hardcopy (6-10 users)
元80,000.00
Electronic (1-5 users)
¥990,000
Electronic (6-10 users)
¥1,406,000
Electronic and 1 Hardcopy (1-5 users)
¥1,140,000
Electronic and 1 Hardcopy (6-10 users)
¥1,556,000
Click here to enquire about additional licenses.
If you are a reseller/distributor please contact us before ordering.
お問合せ、見積および請求書が必要な方はm.murakoshi@idtechex.com までご連絡ください。
HPC hardware market to reach US$581 billion by 2035 at a CAGR of 13.6%.

Report Statistics

Slides 361
Companies Over 50
Forecasts to 2035
Published Dec 2024
 

Preview Content

pdf Document Sample pages
 
 
 
 
ISBN: 9781835700853

Subscription Enquiry