Arm AGI CPU: The world's most efficient agentic CPU
AI Summary
Software agents now reason, decide, and act—placing new demands on the CPU. These workloads require a processor that orchestrates compute, manages accelerators, and coordinates thousands of agents simultaneously. This is a new era of compute, and it demands a new class of processor.
The Arm AGI CPU is the first production silicon from Arm, designed for AI infrastructure at scale. We believe it delivers a new class of CPU with high performance and extreme rack-level density support for agentic AI operations across modern data centers. Based on Arm Neoverse CSS V3, we expect that the addition of the Arm AGI CPU to Arm's data center offering will provide a faster time-to-market through leveraging the same extensive Arm software and hardware ecosystem.
Greater than 2x performance per rack on Arm
The design choices of the Arm AGI CPU are made to deliver maximum performance at rack scale. From microarchitecture to memory, clock frequency to I/O—everything adds up to more performance at gigawatt scale.1
Performance
Modern Arm architecture carries more efficient instruction execution not burdened by decades of legacy complexity. The memory system delivers high bandwidth per core and minimal latency, helping ensure memory does not slow performance.
Scale
Low per-core TDP can support denser deployments and reduced thermal throttling. Each core is dedicated, which can help reduce resource contention and support performance under high thread loads.
Efficiency
High rack density and high performance per watt help ensure maximum utilization of data center space and power resources.
The first partners deploying Arm AGI CPU
Discover how OpenAI, SK Telecom, SAP, Cloudflare, F5, and Cerebras are already using Arm AGI CPU servers in their AI data centers.
Arm AGI CPU servers available now
Arm AGI CPU 1OU Dual NodeReference Server
Reference design for maximum density deployments of Arm AGI CPU – in a OCP DC-MHS standard form factor 1OU Dual Node server.
Arm AGI CPU 2U2PReference Server
19” 2U2P reference design for Arm AGI CPU deployment in a traditional form factor.
Lenovo HR650a V3 2U Arm AGI CPU System
Enterprise-class 2U Arm AGI server optimized for cloud infrastructure, delivering reliable performance and low TCO.
SuperMicro 5U Arm AGI CPU PCIe GPU System
High-density 5U AI platform combining dual Arm AGI CPUs with extensive PCIe GPU expansion.
SuperMicro 2U Hyper Arm AGI CPU Hyper System
Compact 2U dual-socket Arm AGI server designed for efficient cloud and AI infrastructure deployments.
ASRock Rack 2OU2N-ArmSystem
High-density dual-node Arm server built to OCP ORv3 standards for scalable, power-efficient cloud deployments.
Arm AGI CPU specifications and product brief
AI Summary
| Specs | Arm AGI CPU 136C (max core count) | Arm AGI CPU 128C (TCO optimized) | Arm AGI CPU 64C (max mem/core) |
|---|---|---|---|
|
SKU |
|
|
|
|
Processing cores |
|
|
|
|
CPU architecture |
|
|
|
|
System-level cache |
|
|
|
|
Max Frequency |
|
|
|
|
Base TDP* |
|
|
|
|
RDIMM memory |
|
|
|
|
Memory Throughput/core |
|
|
|
|
PCIe/IO |
|
|
|
|
PCIe control lanes |
|
|
|
|
2-Socket support |
|
|
|
|
2 DIMMS per channel |
|
|
|
*Represents a preset TDP value within the configurable TDP range
Talk to an Arm expert to explore how the Arm AGI CPU
is built for next-generation AI data centers.
Powering the next era of AI
Key takeaways
Key takeaways
-
Purpose-built for agentic AI: The first production silicon from Arm is designed for continuous, large-scale AI systems where CPUs orchestrate thousands of parallel tasks across distributed infrastructure.
-
Rack-scale performance and efficiency: Delivers high, sustained performance across dense deployments, with more than 2x performance per rack compared to x86 systems.
-
High-performance cores and memory architecture: Up to 136 Neoverse V3 cores with class-leading 6GB/s memory bandwidth per core at sub-100ns latency.
-
Designed for AI data center efficiency: Operates at 300W TDP, balancing compute density, memory throughput, and system-level efficiency.
Frequently asked questions
Q: What is the Arm AGI CPU and what makes it different from traditional data center CPUs?
A: The Arm AGI CPU is Arm’s first production silicon, designed specifically for agentic AI workloads, delivering high performance, scalable parallel processing, and energy-efficient operation. It enables data centers to run continuous AI workloads at scale while optimizing throughput, resource utilization, and power consumption.
Q: What are the key features of the Arm AGI CPU?
A: The Arm AGI CPU combines high core density, optimized memory architecture, and scalable system design to support AI workloads at scale:
- Efficient cores: Up to 136 Arm Neoverse V3 cores with dedicated 2 MB L2 cache per core and up to 3.7GHz boost frequency, enabling responsive, parallel performance—Arm AGI CPU specs (source required).
- Performance and efficiency: High instruction-per-cycle execution on a TSMC 3 nm process with 300W TDP, balancing compute throughput and energy efficiency—Arm AGI CPU specs (source required).
- Tuned memory architecture: Delivers 6GB/s memory bandwidth per core with support for DDR5-8800 memory and sub-100ns latency, reducing data bottlenecks for AI workloads—Arm AGI CPU specs (source required).
- Built for density: Supports high-density deployments, including reference designs with up to 272 dedicated cores per 1U server and air-cooled configurations compatible with standard infrastructure—Arm AGI CPU specs (source required).
- Rack-scale deployment: Designed for large-scale AI infrastructure, enabling configurations with thousands of cores per rack for continuous AI processing—Arm AGI CPU specs (source required).
- Flexible I/O: Includes 96 PCIe Gen6 lanes, CXL 3.0 for memory expansion, and AMBA CHI links for accelerator connectivity, enabling composable AI systems—Arm AGI CPU specs (source required).
Q: What is agentic AI, and why does it require a new type of CPU?
A: Agentic AI refers to systems that operate continuously, coordinating tasks and making decisions in real time. These workloads require CPUs that can orchestrate distributed systems efficiently at scale.
Q: How does the Arm AGI CPU improve data center performance?
A: It improves performance by delivering high, per-task efficiency and scaling across thousands of cores within a rack, enabling more work per system. This results in more than 2x performance per rack compared to x86 systems.
Q: How does the Arm AGI CPU support AI infrastructure at scale?
A: It supports AI infrastructure by managing distributed workloads, coordinating accelerators, and optimizing data movement across systems to enable continuous, large-scale AI operations.
Q: How does the Arm AGI CPU fit into the Arm ecosystem?
A: It extends the Arm compute platform into production silicon, giving partners the flexibility to deploy Arm technology through IP, compute subsystems, or ready-to-deploy CPUs.
Based on estimates.