데이터 센터 및 클라우드용 AI 칩 2025-2035: 기술 동향, 주요 기업 및 시장 전망

그래픽 처리장치(GPU), 중앙 처리장치(CPU), 맞춤형 AI ASIC, 기타 AI 가속기, 관련 주요 기업에 대한 분석, 기술동향, 공급망 분석 및 시장 전망을 포괄하는 보고서

모두 보기 설명 목차, 표 및 그림 목록 자주 묻는 질문 가격 Related Content
데이터 센터 및 클라우드 시장을 위한 AI 칩, 관련 기술 및 주요 기업들에 대해 분석하고 있는 이번 보고서는 2022년부터 2024년까지의 시장 데이터를 기반으로 현재의 주요 칩 및 새로운 칩 종류(GPU, CPU, 맞춤형 AI ASIC, 기타 AI 칩 등)를 포함하여 40개 이상의 칩 설계업체의 정보를 다루고 있으며, 2025년부터 2035년까지의 시장 예측 및 전망을 제공합니다.
이 보고서에서 아래와 같은 주요 정보를 제공합니다.
 
AI 칩을 위한 하드웨어 평가, 벤치마킹 및 비교
  • 주요 프로세서에 대한 분석 및 벤치마킹, AI 칩의 폼 팩터, 가격 비교, 미국과 중국 주요 기업들의 기술 분석 등 데이터 센터 GPU 에 사용되는 최신 기술 분석
  • 서버용 CPU, 맞춤형 AI ASIC, 기타 AI 칩 등 현재 시장에서 활동 중인 기업들과 신흥 AI 칩의 하드웨어 아키텍처 분석 및 벤치마킹
  • 메모리, 메모리 대역폭, 처리량, 확장성, 가격, 효율성, 고급 공정 노드 등 하드웨어 구성요소 및 현재 AI 칩에 대한 분석, 측정 프레임워크, 역사적 트렌드 분석
  • 프로그래밍 모델, 하드웨어 아키텍처, 첨단 트랜지스터, 첨단 반도체 패키징 등을 포함하여 AI 칩 설계 및 제조에 필요한 주요 요소 분석
 
시장 정보
  • 주요 기술 유형별 시장 성장을 이끄는 요인, 저해 요인 및 향후 전망
  • 첨단 집적 회로를 위한 다양한 반도체 제조사 및 칩 설계자의 역량을 포함한 AI 칩 공급망 분석
  • 정부 투자, 첨단 반도체 패키징 공장에 대한 투자, 하이퍼스케일러의 자본 지출(CapEx), 칩 설계자의 수익 등을 포함한 투자 정보
  • 수출 허가가 필요한 칩 유형 분석 포함한 2022년 이후 미국의 대외 반도체 수출 정책 분석
 
시장 전망 및 분석
  • 주요 AI 칩 기술 유형별로 세분화된 10년 시장 전망
  • AI 데이터 센터 및 클라우드 인프라에 사용되는 AI 칩에 대한 주요 기술 및 상업적 트렌드 평가
 
이 보고서에서 다루는 주요 내용/목차는 아래와 같습니다.
 
1. 핵심 요약
2 AI 모델 및 AI 칩 소개
2.1 AI의 기초
2.2 대규모 AI 모델
3 기술 개요
3.1 명령어 집합 아키텍처
3.2 프로그래밍 모델
3.3 하드웨어 아키텍처
3.4 트랜지스터
3.5 첨단 반도체 패키징
4 AI 기능을 갖춘 중앙 처리 장치(CPU)
4.1 CPU의 기술 개요
4.2 주요 기업 현황
5 그래픽 처리 장치(GPU)
5.1 GPU의 시장 개요
5.2 GPU 기술 분석
5.3 주요 기업 현황
6 클라우드 서비스 제공업체(CSP)를 위한 맞춤형 AI ASIC
6.1 CSP용 맞춤형 AI ASIC의 시장 개요
6.2 CSP용 맞춤형 AI ASIC의 하드웨어 분석
6.3 주요 기업 현황
7 기타 AI 칩
7.1 이기종 매트릭스 기반 AI 가속기 및 주요 기업 현황
7.2 공간 기반 AI 가속기와 주요 기업 현황
7.3 조립형 재구성 가능 어레이(CGRA)
8 벤치마크 및 하드웨어 트렌드
8.1 AI 칩 벤치마킹
8.2 성능 및 확장성
8.3 에너지 효율성
8.4 메모리 및 메모리 대역폭
8.5 새로운 AI 가속기의 성능 평가를 위한 고려 사항
9 공급망, 투자 및 무역 규제
9.1 공급망
9.2 투자
9.3 무역 규제
10 시장 전망
 
Frontier AI attracts hundreds of billions in global investment, with governments and hyperscalers racing to lead in domains like drug discovery and autonomous infrastructure. Graphics processing units (GPUs) and other AI chips have been instrumental in driving this growth of artificial intelligence, providing the compute needed for deep learning within data centers and cloud infrastructure. GPUs have been pivotal in delivering computational capabilities, being the dominant undercurrent below the wave that is large language models (LLMs) and generative AI. However, with the demand for more efficient computation, lower costs, higher performance, massively scalable systems, faster inference, and domain-specific computation, there is opportunity for other AI chips to grow in popularity.
 
As the landscape of AI chips broadens past just GPUs, with novel architectures reaching widescale commercialization, IDTechEx's report "AI Chips for Data Centers and Cloud 2025-2035: Technologies, Market, Forecasts" offers an independent analysis of the AI chip market for data centers and the cloud. This includes benchmarking current and emerging technologies, technology breakdowns, and key trends, covering current and emerging hardware architectures, advanced node technologies, and advanced semiconductor packaging, as well as information on supply chain, investments, and policy. Granular revenue forecasts from 2025 to 2035 of the data center and cloud AI chips market are provided, segmented by types of AI chips. These include GPUs, custom AI application-specific integrated circuits (ASICs) used by hyperscalers and cloud service providers (CSPs), AI-capable central processing units (CPUs), and other AI ASICs developed by both AI chip-focused startups and large vendors.
 
 
Graphics Processing Units (GPUs)
The largest systems for AI are massive scale-out HPC and AI systems - these heavily implement GPUs. These tend to be hyperscaler AI data centers and supercomputers, both of which can offer exaFLOPS of performance, on-premise or over distributed networks. NVIDIA has seen remarkable success over recent years with its Hopper (H100/H200) chips and recently released Blackwell (B200/B300) chips. AMD has also created competitive chips with its MI300 series processors (MI300X/MI325X). Chinese players are also developing solutions due to sanctions from the US on advanced chips. These high-performance GPUs continue to adopt the most advanced semiconductor technologies. One example is increased on-chip memory capacity, with top chips having over 250GB of high-bandwidth memory (HBM), enabling larger AI models with even more parameters to run on these GPUs. These chips also adopt the most advanced semiconductor packaging solutions, such as TSMC's CoWoS-L packaging, chiplets, and multi-die GPUs, as well as the most advanced process nodes (5nm and below). All of these trends and market activities are explored in detail in this report.
 
Custom AI Chips Used by Hyperscalers and Cloud Service Providers
GPUs have been fundamental for training AI models but face limitations, such as high total cost of ownership (TCO), vendor lock-in risks, low utilization for AI-specific operations, and can be overkill for specific inference workloads. An emerging strategy that hyperscalers are adopting is using systolic array-based custom AI ASICs. These have purpose-built cores for AI workloads, are cheaper per operation, are specialized for particular systems (e.g., transformers, recommender systems, etc), offer efficient inference, and give hyperscalers and CSPs the opportunity for full-stack control and differentiation without sacrificing performance. Evaluation of potential risks, key partnerships, player activity, benchmarking, and technology overviews is available with this report.
 
Other AI Chips
Other AI chips are being commercialized to disrupt GPUs, with both similar and novel computing architectures. Some large chip vendors, such as Intel, Huawei, and Qualcomm, have designed AI accelerators (e.g., Gaudi, Ascend 910, Cloud AI 100), using heterogeneous arrays of compute units (similar to GPUs), but purpose-built to accelerate AI workloads. These offer a balance between performance, power efficiency, and flexibility for specific application domains. Often, these chips will contain matrix engines and tensor cores, which are designed to execute dense linear algebra operations like GEMM (General Matrix Multiply) and BMM (Batch Matrix Multiply) with high throughputs.
 
AI chip-focused startups often take a different approach, deploying cutting-edge architectures and fabrication techniques with the likes of dataflow-controlled processors, wafer-scale packaging, spatial AI accelerators, processing-in-memory (PIM) technologies, and coarse-grained reconfigurable arrays (CGRAs). Various companies have successfully launched these systems (Cerebras, Groq, Graphcore, SambaNova, Untether AI, and others) for data centers and cloud computing, often developing rack-scale solutions for easy enterprise deployment or offering usage on their own cloud platforms. These systems perform exceptionally, especially in scale-up environments. IDTechEx's report offers comprehensive benchmarking, comparisons, key trends, technology breakdowns, and player activity.
 
Designing AI chips and supply chain
Developing an AI chip with competitive throughput for training (time-to-train) and inference (tokens per second), high-energy efficiencies (TOPS/watt), and associated software support is a stringent challenge for all chip designers. This process involves a fine balance of many steps, including selecting programming and execution models, designing optimized hardware and memory architecture, and fabrication with advanced process nodes and advanced semiconductor packaging. For instance, data center chips are adopting the most advanced process nodes from TSMC, Intel Foundry, and Samsung Foundry, using EUV (extreme ultraviolet) lithography techniques from ASML. These foundries are pushing transistor technologies past 5nm technologies using FinFET (Fin-field effect transistor) to sub-2nm nodes using GAAFET (gate-all-around FETs) with backside power delivery. Recent fabrication developments, device requirements, hardware architecture breakdowns, advanced semiconductor packaging details, supply chain, and programming model comparisons are all included throughout this report.
 
The various technologies involved in designing and manufacturing give wide breadth for future technological innovation across the semiconductor industry supply chain. Government policy and heavy investment show the prevalent interest in pushing frontier AI toward new heights, and this will require exceptional volumes of AI chips within AI data centers to meet this demand. IDTechEx forecasts this market will grow at a CAGR of 14% from 2025 to 2030, with revenues exceeding US$400 billion.
Key Aspects
Hardware evaluation, benchmarking, and comparison for AI chips
  • Exploring current technologies used in data center GPUs, including analysis and benchmarking of leading processors, AI chip form factors, pricing comparisons, and technology breakdowns for leading US and Chinese players.
  • Detailing and benchmarking hardware architectures for current market players and emerging AI chips for server CPUs, custom AI ASICs, and other AI chips, including AI accelerators using heterogeneous matrix-based systems and spatial AI accelerators.
  • Analysis, measurement frameworks, and historical trends of hardware components and current AI chips, including memory, memory bandwidth, throughput (and other performance metrics), scalability, pricing, efficiency, and advanced process nodes.
  • Breakdown of key elements to designing and manufacturing AI chips, including programming models, hardware architectures, advanced transistors, and advanced semiconductor packaging.
 
Market information
  • Commentary detailing drivers and barriers of key technology types, as well as expected outlooks.
  • Analysis of the AI chip supply chain, including capabilities of various semiconductor manufacturers and chip designers for advanced integrated circuits.
  • Investment information, including governmental investments, investments into advanced semiconductor packaging plants, hyperscaler capex, and chip designer revenues.
  • Breakdown of US policy since 2022 concerning the export of US chips to foreign nations, including analysis of which chips require export licenses.
 
Market forecast and analysis
  • 10-year granular market forecasts separated by key AI chip technology types.
  • Assessment of key technological and commercial trends for AI chips used for AI data centers and cloud infrastructure.
Report MetricsDetails
Historic Data2022 - 2024
CAGRThe total market for AI chips to reach US$473.2 billion by 2035 at a CAGR of 7.2% over the forecast period
Forecast Period2025 - 2035
Forecast UnitsRevenue (US$ Billions)
Regions CoveredWorldwide
Segments CoveredServer central processing units (CPUs) with AI extensions, data center graphics processing units (GPUs), custom AI ASICs used by CSPs and hyperscalers, other data center AI chips (ASICs from AI Chip-focused startups, ASICs from large vendors)
IDTechEx의 분석가 액세스
모든 보고서 구입에는 전문가 분석가와의 최대 30분의 전화통화 시간이 포함되어, 보고서의 주요 결과를 귀하가 제시하는 비즈니스 문제에 연결하도록 돕습니다. 이 전화통화는 보고서를 구매한 후 3개월 이내에 사용해야합니다.
추가 정보
이 보고서에 대해 궁금한 점이 있으시면 언제든지 research@IDTechEx.com으로 보고서 팀에 문의하거나, 영업 관리자에게 문의하십시오

AMERICAS (USA): +1 617 577 7890
ASIA (Japan): +81 3 3216 7209
ASIA (Korea): +82 10 3896 6219
EUROPE (UK) +44 1223 812300
1.EXECUTIVE SUMMARY
1.1.What is AI?
1.2.What are AI chips for data center and cloud?
1.3.AI chips must improve as performance of AI models outstrips Moore's Law
1.4.Large AI models require scaling of more AI chips
1.5.Market dynamics and strategic shifts in AI hardware
1.6.Layers to designing an AI chip
1.7.Types of AI Chips
1.8.Technology readiness of AI chip technologies
1.9.AI chip technologies benchmarked
1.10.AI chip landscape - Chip designers
1.11.Graphics Processing Units (GPUs)
1.12.Trends in high-performance data center GPUs
1.13.ASICs used by major cloud service providers for accelerating AI workloads
1.14.Trends in GPU alternatives for AI data center
1.15.AI chip key workloads for training and inference
1.16.Hardware demands for training and inference
1.17.Inference benchmarks show real time performance of top GPUs
1.18.Performance of common AI chips: FP16/BF16 precisions
1.19.Trends in advanced process nodes and energy efficiency in the last decade
1.20.Key players: AI chip supply chain
1.21.Government industrial policy and funding for semiconductor industry
1.22.US Sanctions on AI chips to China
1.23.Market size forecast of AI chips: 2025-2035
1.24.Annotated market size forecast of GPUs: 2025-2035
1.25.Drivers and challenges for AI chip adoption
1.26.Access More With an IDTechEx Subscription
2.INTRODUCTION TO AI MODELS AND AI CHIPS
2.1.1.What is AI?
2.1.2.What is an AI chip?
2.1.3.AI acceleration
2.1.4.Types of AI chip product categories
2.1.5.Overview of major AI chip markets
2.1.6.Cloud and data center computing
2.1.7.Users, procurement and partnerships of cloud and data center compute
2.1.8.Cloud AI
2.1.9.Enterprise core
2.1.10.Telecom edge
2.1.11.Edge vs Cloud characteristics
2.1.12.Key players: AI chip supply chain
2.2.Fundamentals of AI
2.2.1.Fundamentals of AI: Algorithms, Data, and Hardware
2.2.2.Training and inference
2.2.3.AI chips use low-precision computing
2.2.4.Common number representations in AI chips
2.2.5.Parallel computing: Data parallelism and model parallelism
2.2.6.Deep learning: how an AI algorithm is implemented
2.2.7.Neural networks explained
2.2.8.Types of Neural Networks
2.2.9.Types of neural networks and use cases
2.3.Large AI Models
2.3.1.Notable AI models increasing performance at a rate of 4.5x a year since 2010
2.3.2.Transformers used for LLM replace RNNs for natural language processing
2.3.3.Language, computer vision, and multimodal AI models are the most popular
2.3.4.Reasons for AI performance outpacing Moore's Law
2.3.5.Key drivers for continued growth of AI models
2.3.6.Scale-up and scale-out systems
2.3.7.Training AI models is very energy intensive
2.3.8.Hardware design and energy inefficiencies of compute
2.3.9.MLPerf Power: Power ranges for various AI chip types and applications
3.TECHNOLOGY OVERVIEW
3.1.AI Chips Hardware Design Overview
3.1.1.History of computer hardware
3.1.2.Progression of AI hardware
3.1.3.Trends in AI chips to expect
3.1.4.Layers to designing an AI chip
3.2.Instruction Set Architectures
3.2.1.Introduction to Instruction Set Architectures (ISAs) for AI workloads
3.2.2.CISC and RISC ISAs for AI accelerators
3.3.Programming Models and Execution Models
3.3.1.Programming model vs execution model
3.3.2.Flynn's taxonomy and programming models
3.3.3.Important execution models and programming models for AI chips
3.3.4.Introduction to Von Neumann Architecture
3.3.5.Von Neumann compared with common programming models
3.4.Hardware Architectures
3.4.1.ASICs, FPGAs, and GPUs used for neural network architectures
3.4.2.Benchmarking capabilities of AI chips
3.4.3.Types of AI Chips
3.4.4.TRL of AI chip technologies
3.4.5.Pros and cons of commercial AI chips
3.4.6.Pros and cons of emerging AI chips
3.4.7.Technologies found in general-purpose processors
3.4.8.Special-purpose resources
3.4.9.Accelerator taxonomy
3.5.Transistors
3.5.1.How transistors operate: p-n junctions
3.5.2.Moore's law
3.5.3.Gate length reductions pose challenges to planar FETs below 20nm
3.5.4.Increasing Transistor Count
3.5.5.Planar FET to FinFET
3.5.6.GAAFET, MBCFET, RibbonFET
3.5.7.TSMC's leading-edge nodes roadmap
3.5.8.Intel Foundry's leading-edge nodes roadmap
3.5.9.Samsung Foundry's leading-edge nodes roadmap
3.5.10.CFETs to be used beyond GAAFET scaling
3.5.11.Device architecture roadmap (I)
3.5.12.Scaling technology roadmap overview
3.6.Advanced Semiconductor Packaging
3.6.1.Progression from 1D to 3D semiconductor packaging
3.6.2.Key metrics for advanced semiconductor packaging performance
3.6.3.Overview of interconnection technique in semiconductor packaging
3.6.4.Overview of 2.5D packaging structure
3.6.5.2.5D advanced semiconductor packaging technology portfolio
3.6.6.2.5D advanced semiconductor packaging used in top AI chips
3.6.7.Overcoming die size limitations
3.6.8.Integrated heterogeneous systems
3.6.9.Case study: AMD MI300A CPU/GPU heterogenous integration
3.6.10.Future system-in-package architecture
3.6.11.For more information on advanced semiconductor packaging
4.AI-CAPABLE CENTRAL PROCESSING UNITS (CPUS)
4.1.Technology Overview of CPUs
4.1.1.CPU introduction
4.1.2.Core architecture of a HPC and AI CPU
4.1.3.Key CPU requirements for HPC and AI workloads (1)
4.1.4.Key CPU Requirements for HPC and AI Workloads (2)
4.1.5.AVX-512 vector extensions for x86-64 Instruction Set
4.2.Intel CPUs
4.2.1.Intel: Xeon CPUs for data center
4.2.2.Intel: Advanced Matrix Extensions in CPUs for built-in AI acceleration
4.2.3.Intel: 4th Gen Xeon Scalable Processor performance with AMX
4.3.AMD CPUs
4.3.1.AMD: EPYC CPUs for data center
4.4.IBM CPUs
4.4.1.IBM: Power CPUs for data center
4.5.Arm CPUs
4.5.1.Arm licenses core designs with its RISC-based ISAs
4.5.2.Arm CPUs for data center
4.5.3.CPU outlook
5.GRAPHICS PROCESSING UNITS (GPUS)
5.1.Market Overview of GPUs
5.1.1.Types of AI GPUs
5.1.2.Historical background of GPUs
5.1.3.GPUs popularity since the 2010s
5.1.4.Data center GPU player landscape by region
5.1.5.Commercial activity of key US and Chinese data center GPU manufacturers
5.1.6.Drivers: Technology advancements and market opportunities
5.1.7.Drivers: Energy efficiency, performance, incentives, and brand strength
5.1.8.Barriers: Monopolization, competition, and product complexity
5.1.9.Barriers: R&D, competition from customers, market consolidation
5.1.10.How can startups compete with GPU market leaders
5.2.GPU Technology Breakdown
5.2.1.Key architectural differences between CPUs and GPUs
5.2.2.Architecture breakdown of high-performance data center GPUs
5.2.3.Data center GPUs key features
5.2.4.NVIDIA and AMD data center GPUs benchmark
5.2.5.Consumer GPUs as cloud compute
5.2.6.Workstation / professional GPUs as cloud compute
5.2.7.Pricing of GPUs by type
5.2.8.Form factor options for GPUs
5.2.9.Pricing of data center GPU form factors
5.2.10.Threads show how latency and throughput is handled by GPUs and CPUs
5.2.11.NVIDIA and AMD software
5.2.12.Trends in high-performance data center GPUs
5.2.13.Trends in high-performance data center GPUs
5.3.NVIDIA GPUs
5.3.1.NVIDIA: Tensor mathematics
5.3.2.NVIDIA: Tensor cores
5.3.3.NVIDIA: NVIDIA CUDA and tensor cores
5.3.4.NVIDIA: Data center GPU product timeline
5.3.5.NVIDIA: Ampere GPUs
5.3.6.NVIDIA: Hopper GPUs
5.3.7.NVIDIA: Blackwell GPUs (I)
5.3.8.NVIDIA: Blackwell GPU (II)
5.3.9.NVIDIA: Rack-scale solutions
5.4.AMD GPUs
5.4.1.AMD: CDNA 3 Architecture and Compute Units for GPU Compute
5.4.2.AMD: MI325X GPU
5.4.3.AMD: Instinct GPU and competitive positioning
5.4.4.AMD: MI300A CPU/GPU memory coherency with heterogenous integration
5.5.Intel GPUs
5.5.1.Intel: Intel GPU Max and the Xe-HPC Architecture
5.5.2.Intel: Future ASIC and general-purpose GPU
5.6.Chinese GPUs
5.6.1.Biren Technologies: Chinese GPGPU
5.6.2.Biren Technologies: BR100 and BR104 Chinese GPGPU
5.6.3.Moore Threads: MTT S4000 Chinese GPU
5.6.4.MetaX: MXC500 Chinese GPGPU
5.6.5.Iluvatar CoreX: Tianyuan 100 and Zhikai 100 Chinese GPGPUs
5.6.6.GPU Outlook
6.CUSTOM AI ASICS FOR CLOUD SERVICE PROVIDERS (CSPS)
6.1.Market Overview of Custom AI ASICs for CSPs
6.1.1.Introduction to custom application-specific integrated circuits (ASICs)
6.1.2.AI ASICs based on application
6.1.3.Custom ASICs enter the market to compete with GPUs
6.1.4.Drivers for investment, and challenges for custom ASICs
6.1.5.CSP custom ASIC player landscape by region
6.1.6.ASICs used by major cloud service providers for accelerating AI workloads
6.1.7.AI ASIC companies' capabilities
6.2.Hardware Breakdown of Custom AI ASICs for CSPs
6.2.1.GPU and ASIC comparison
6.2.2.Cloud service provider ASICs have similar architectures, using systolic arrays
6.2.3.Systolic arrays in ASICS are an alternative to tensor cores in GPUs
6.2.4."Systolic array lock-in"
6.3.Key Players
6.3.1.Google TPU
6.3.2.Amazon: Trainium and Inferentia
6.3.3.Amazon: Trainium and Inferentia chip components and packaging
6.3.4.Microsoft: Maia
6.3.5.Meta: MTIA
6.3.6.Future US ASIC players
6.3.7.Chinese ASIC players and Chinese AI chips from cloud service providers
6.3.8.Outlook
7.OTHER AI CHIPS
7.1.1.Introduction to other architectures: Chapter Overview
7.1.2.Other AI chips player landscape by region
7.2.Heterogenous Matrix-Based AI Accelerators
7.2.1.Heterogenous matrix-based AI accelerators
7.2.2.Heterogenous matrix-based AI accelerators architectures
7.2.3.Habana: Gaudi
7.2.4.Intel: Gaudi2
7.2.5.Intel: Greco
7.2.6.Intel: Gaudi3
7.2.7.Cambricon Technologies: Siyuan 370 is China's AI tensor-based AI chip
7.2.8.Huawei: Ascend 910
7.2.9.Huawei: Da Vinci architecture
7.2.10.Baidu: Kunlun and XPU
7.2.11.Qualcomm: Cloud AI 100
7.2.12.Qualcomm: AI core
7.2.13.Summary of key players
7.3.Spatial AI Accelerators
7.3.1.Spatial AI accelerators
7.3.2.Cerebras: Wafer-scale processors as a competitor to GPUs
7.3.3.Cerebras: WSE-3
7.3.4.SambaNova: Reconfigurable dataflow processors as substitute to GPUs
7.3.5.SambaNova: SN40L Reconfigurable Dataflow Unit (RDU)
7.3.6.Graphcore: Second-generation Colossus™ MK2 IPU processor
7.3.7.Graphcore: Bow IPU and Pods
7.3.8.Groq: Natural language processor designed for AI inference
7.3.9.Groq: Performance and technology
7.3.10.Untether AI: SpeedAI240 uses at-memory computation
7.3.11.Key players summary (I)
7.3.12.Key players summary (II)
7.4.Coarse-Grained Reconfigurable Arrays (CGRAs)
7.4.1.CGRAs could be a future contender for mainstream compute fabrics
7.4.2.CGRA breakdown
7.4.3.Future outlook - the search for flexible architectures with high energy efficiency and performance
7.4.4.CGRAs vs dataflow vs manycore
7.4.5.Trends in GPU alternatives for AI data center
7.4.6.Trends in other AI chips
8.BENCHMARKS AND HARDWARE TRENDS
8.1.Benchmarking AI Chips
8.1.1.MLPerf by MLCommons for benchmarking AI chips
8.1.2.MLCommons benchmarks: Training and inference key workloads and models
8.1.3.AI chip capabilities (I)
8.1.4.AI chip capabilities (II)
8.1.5.Training benchmarking
8.1.6.Inference benchmarking
8.1.7.AI chip technologies benchmarked
8.2.Performance and Scalability
8.2.1.MLPerf Inference: Data Center: Tokens per second
8.2.2.MLPerf Training: Natural Language Processing performance
8.2.3.MLPerf Training: NVIDIA performance
8.2.4.MLPerf Training: Scalability of Google TPUs
8.2.5.NVIDIA and AMD data center GPU throughput with OpenCL benchmark
8.2.6.Neocloud giants: GPU inference performance and GPU scalability
8.2.7.Performance of common AI chips: FP16/BF16 precisions
8.2.8.Performance of common AI chips: Comparing different precisions
8.3.Energy Efficiency
8.3.1.Performance per watt for different AI chips
8.3.2.Trends in advanced process nodes and energy efficiency in the last decade
8.4.Memory and Memory Bandwidth
8.4.1.Key challenge: The memory wall
8.4.2.Illustrating the memory wall: Memory hierarchy latency bottleneck
8.4.3.Memory bandwidths of different chip types
8.4.4.High bandwidth memory (HBM) and comparison with other DRAM technologies
8.4.5.Evolution of HBM generations and transition to HBM4
8.4.6.Benchmarking of HBM technologies in the market from key players (1)
8.4.7.Benchmarking of HBM technologies in the market from key players (2)
8.4.8.Memory bandwidth trends
8.4.9.Memory capacity trends
8.5.Considerations for Evaluating Performance of New AI Accelerators
8.5.1.Evaluating performance of AI accelerators
8.5.2.Performance of accelerators must be measured across various metrics
8.5.3.Latency must be optimized through various strategies
8.5.4.Fundamentals abundant data computing systems using the Roofline Model
8.5.5.Peak throughput is limited by DNN accelerator design constraints
8.5.6.Hardware design and energy inefficiencies of compute
8.5.7.Flexibility is key for handling wide range of DNNs
8.5.8.Network on Chip - example from academia showing flexibility
9.SUPPLY CHAIN, INVESTMENTS, AND TRADE RESTRICTIONS
9.1.Supply Chain
9.1.1.IC supply chain player categories
9.1.2.Integrated circuit supply chain models
9.1.3.Supply chain by production process
9.1.4.Concentration of AI chip supply chain
9.1.5.Populated supply chain for AI chips
9.1.6.Populated supply chain for AI chips by component
9.1.7.AI chip landscape - Chip designers
9.1.8.Populated supply chain for custom integrated circuits
9.1.9.IDM fabrication capabilities
9.1.10.Foundry capabilities
9.1.11.AI cloud categories and players
9.1.12.US hyperscalers capital expenditure
9.2.Investments
9.2.1.Government industrial policy and funding for semiconductor industry
9.2.2.Government investments in US and European advanced packaging
9.2.3.Government investments in Asian packaging and the TMSC supply chain
9.3.Trade Restrictions
9.3.1.US policy regarding advanced semiconductors in China and other nations
9.3.2.Oct 7th, 2022 US sanctions on China technologies
9.3.3.Oct 17th, 2023, US sanctions on AI chips (I)
9.3.4.Oct 17th, 2023, US sanctions on AI chips (II)
9.3.5.AI chips compliant in China
9.3.6.Dec 2nd, 2024, further controls on advanced computing and semiconductor manufacture
9.3.7.Restrictions on High-Bandwidth Memory (HBM)
9.3.8.Jan 13th, 2025, AI Diffusion Framework (US worldwide export controls) (I)
9.3.9.Jan 13th, 2025, AI Diffusion Framework (US worldwide export controls) (II)
9.3.10.NVIDIA revenues by geography, affected by US restrictions
10.FORECASTS
10.1.1.Forecast methodology
10.1.2.Forecast assumptions and outlook
10.1.3.Market size forecast of AI chips: 2025-2035
10.1.4.Market share forecast of AI chips: 2025-2035
10.1.5.Annotated market size forecast of GPUs: 2025-2035
10.1.6.IDTechEx outlook for GPUs
10.1.7.Custom AI ASIC market value
10.1.8.Annotated market size forecast of custom AI ASICs: 2025-2035
10.1.9.IDTechEx outlook for custom AI ASIC chips
10.1.10.Annotated market size forecast of other AI chips: 2025-2035
10.1.11.IDTechEx outlook for other AI chip architectures
 

About IDTechEx reports

What are the qualifications of the people conducting IDTechEx research?

Content produced by IDTechEx is researched and written by our technical analysts, each with a PhD or master's degree in their specialist field, and all of whom are employees. All our analysts are well-connected in their fields, intensively covering their sectors, revealing hard-to-find information you can trust.

How does IDTechEx gather data for its reports?

By directly interviewing and profiling companies across the supply chain. IDTechEx analysts interview companies by engaging directly with senior management and technology development executives across the supply chain, leading to revealing insights that may otherwise be inaccessible.
 
Further, as a global team, we travel extensively to industry events and companies to conduct in-depth, face-to-face interviews. We also engage with industry associations and follow public company filings as secondary sources. We conduct patent analysis and track regulatory changes and incentives. We consistently build on our decades-long research of emerging technologies.
 
We assess emerging technologies against existing solutions, evaluate market demand and provide data-driven forecasts based on our models. This provides a clear, unbiased outlook on the future of each technology or industry that we cover.

What is your forecast methodology?

We take into account the following information and data points where relevant to create our forecasts:
  • Historic data, based on our own databases of products, companies' sales data, information from associations, company reports and validation of our prior market figures with companies in the industry.
  • Current and announced manufacturing capacities
  • Company production targets
  • Direct input from companies as we interview them as to their growth expectations, moderated by our analysts
  • Planned or active government incentives and regulations
  • Assessment of the capabilities and price of the technology based on our benchmarking over the forecast period, versus that of competitive solutions
  • Teardown data (e.g. to assess volume of materials used)
  • From a top-down view: the total addressable market
  • Forecasts can be based on an s-curve methodology where appropriate, taking into account the above factors
  • Key assumptions and discussion of what can impact the forecast are covered in the report.

How can I be confident about the quality of work in IDTechEx reports?

Based on our technical analysts and their research methodology, for over 25 years our work has regularly received superb feedback from our global clients. Our research business has grown year-on-year.
 
Recent customer feedback includes:
"It's my first go-to platform"
- Dr. Didi Xu, Head of Foresight - Future Technologies, Freudenberg Technology Innovation
 
"Their expertise allows us to make data-driven, strategic decisions and ensures we remain aligned with the latest trends and opportunities in the market."
- Ralf Hug, Global Head of Product Management & Marketing, Marquardt

What differentiates IDTechEx reports?

Our team of in-house technical analysts immerse themselves in industries over many years, building deep expertise and engaging directly with key industry players to uncover hard-to-find insights. We appraise technologies in the landscape of competitive solutions and then assess their market demand based on voice-of-the-customer feedback, all from an impartial point of view. This approach delivers exceptional value to our customers—providing high-quality independent content while saving customers time, resources, and money.

Why should we pick IDTechEx research over AI research?

A crucial value of IDTechEx research is that it provides information, assessments and forecasts based on interviews with key people in the industry, assessed by technical experts. AI is trained only on content publicly available on the web, which may not be reliable, in depth, nor contain the latest insights based on the experience of those actively involved in a technology or industry, despite the confident prose.

How can I justify the ROI of this report?

Consider the cost of the IDTechEx report versus the time and resources required to gather the same quality of insights yourself. IDTechEx analysts have built up an extensive contact network over many years; we invest in attending key events and interviewing companies around the world; and our analysts are trained in appraising technologies and markets.
 
Each report provides an independent, expert-led technical and market appraisal, giving you access to actionable information immediately, rather than you having to spend months or years on your own market research.

Can I speak to analysts about the report content?

All report purchases include up to 30 minutes of telephone time with an expert analyst who will help you link key findings in the report to the business issues you're addressing. This needs to be used within three months of purchasing the report.

What is the difference between a report and subscription?

A subscription from IDTechEx can include more reports, access to an online information platform with continuously updated information from our analysts, and access to analysts directly.

Before purchasing, I have some questions about the report, can I speak to someone?

Please email research@idtechex.com stating your location and we will quickly respond.

About IDTechEx

Who are IDTechEx's customers?

IDTechEx has served over 35,000 customers globally. These range from large corporations to ambitious start-ups, and from Governments to research centers. Our customers use our work to make informed decisions and save time and resources.

Where is IDTechEx established?

IDTechEx was established in 1999, and is headquartered in Cambridge, UK. Since then, the company has significantly expanded and operates globally, having served customers in over 80 countries. Subsidiary companies are based in the USA, Germany and Japan.

Questions about purchasing a report

How do I pay?

In most locations reports can be purchased by credit card, or else by direct bank payment.

How and when do I receive access to IDTechEx reports?

When paying successfully by credit card, reports can be accessed immediately. For new customers, when paying by bank transfer, reports will usually be released when the payment is received. Report access will be notified by email.

How do I assign additional users to the report?

Users can be assigned in the report ordering process, or at a later time by email.

Can I speak to someone about purchasing a report?

Please email research@idtechex.com stating your location and we will quickly respond.
 

Ordering Information

데이터 센터 및 클라우드용 AI 칩 2025-2035: 기술 동향, 주요 기업 및 시장 전망

£$¥
전자 (사용자 1-5명)
£5,650.00
전자 (사용자 6-10명)
£8,050.00
전자 및 1 하드 카피 (사용자 1-5명)
£6,450.00
전자 및 1 하드 카피 (사용자 6-10명)
£8,850.00
전자 (사용자 1-5명)
€6,400.00
전자 (사용자 6-10명)
€9,100.00
전자 및 1 하드 카피 (사용자 1-5명)
€7,310.00
전자 및 1 하드 카피 (사용자 6-10명)
€10,010.00
전자 (사용자 1-5명)
$7,000.00
전자 (사용자 6-10명)
$10,000.00
전자 및 1 하드 카피 (사용자 1-5명)
$7,975.00
전자 및 1 하드 카피 (사용자 6-10명)
$10,975.00
전자 (사용자 1-5명)
元50,000.00
전자 (사용자 6-10명)
元72,000.00
전자 및 1 하드 카피 (사용자 1-5명)
元58,000.00
전자 및 1 하드 카피 (사용자 6-10명)
元80,000.00
전자 (사용자 1-5명)
¥990,000
전자 (사용자 6-10명)
¥1,406,000
전자 및 1 하드 카피 (사용자 1-5명)
¥1,140,000
전자 및 1 하드 카피 (사용자 6-10명)
¥1,556,000
전자 (사용자 1-5명)
₩9,800,000
전자 (사용자 6-10명)
₩14,000,000
전자 및 1 하드 카피 (사용자 1-5명)
₩11,200,000
전자 및 1 하드 카피 (사용자 6-10명)
₩15,400,000
Click here to enquire about additional licenses.
If you are a reseller/distributor please contact us before ordering.
お問合せ、見積および請求書が必要な方はm.murakoshi@idtechex.com までご連絡ください。
인공지능 반도체 (AI Chip) 시장은 2025년부터 연평균 14%의 성장률로 2030년에는 4,530억 달러 규모의 시장을 형성할 것으로 전망

보고서 통계

슬라이드 333
전망 2035
게시 Apr 2025
 

콘텐츠 미리보기

pdf Document Sample pages
 

Customer Testimonial

quote graphic
"The resources provided by IDTechEx, such as their insightful reports and analysis, engaging webinars, and knowledgeable analysts, serve as valuable tools and information sources... Their expertise allows us to make data-driven, strategic decisions and ensures we remain aligned with the latest trends and opportunities in the market."
Global Head of Product Management and Marketing
Marquardt GmbH
 
 
 
ISBN: 9781835701201

Subscription Enquiry