Overview

Arm AGI CPU: The world's most efficient agentic CPU

AI Summary

Software agents now reason, decide, and act—placing new demands on the CPU. These workloads require a processor that orchestrates compute, manages accelerators, and coordinates thousands of agents simultaneously. This is a new era of compute, and it demands a new class of processor.

 

The Arm AGI CPU is the first production silicon from Arm, designed for AI infrastructure at scale. We believe it delivers a new class of CPU with high performance and extreme rack-level density support for agentic AI operations across modern data centers. Based on Arm Neoverse CSS V3, we expect that the addition of the Arm AGI CPU to Arm's data center offering will provide a faster time-to-market through leveraging the same extensive Arm software and hardware ecosystem.

Features

Key features of the Arm AGI CPU

Rack-level performance

Performance

Greater than 2x performance per rack on Arm

The design choices of the Arm AGI CPU are made to deliver maximum performance at rack scale. From microarchitecture to memory, clock frequency to I/O—everything adds up to more performance at gigawatt scale.1

Performance

Modern Arm architecture carries more efficient instruction execution not burdened by decades of legacy complexity. The memory system delivers high bandwidth per core and minimal latency, helping ensure memory does not slow performance.

Single tall server rack with visible components

Scale

Low per-core TDP can support denser deployments and reduced thermal throttling. Each core is dedicated, which can help reduce resource contention and support performance under high thread loads.

Single tall server rack with visible components

Efficiency

High rack density and high performance per watt help ensure maximum utilization of data center space and power resources.

Single tall server rack with visible components
Single tall server rack with visible components

The first partners deploying Arm AGI CPU

Discover how OpenAI, SK Telecom, SAP, Cloudflare, F5, and Cerebras are already using Arm AGI CPU servers in their AI data centers.

 
More from our partners
SERVERS

Arm AGI CPU servers available now  

Arm logo

Arm AGI CPU 1OU Dual Node
Reference Server

Reference design for maximum density deployments of Arm AGI CPU – in a OCP DC-MHS standard form factor 1OU Dual Node server.

Arm logo

Arm AGI CPU 2U2P
Reference Server

19” 2U2P reference design for Arm AGI CPU deployment in a traditional form factor.

Lenovo logo

Lenovo HR650a V3 2U Arm AGI CPU System

Enterprise-class 2U Arm AGI server optimized for cloud infrastructure, delivering reliable performance and low TCO.

Supermicro-logo

SuperMicro 5U Arm AGI CPU PCIe GPU System

High-density 5U AI platform combining dual Arm AGI CPUs with extensive PCIe GPU expansion.

Supermicro-logo

SuperMicro 2U Hyper Arm AGI CPU Hyper System

Compact 2U dual-socket Arm AGI server designed for efficient cloud and AI infrastructure deployments.

Asrock-logo

ASRock Rack 2OU2N-Arm
System

High-density dual-node Arm server built to OCP ORv3 standards for scalable, power-efficient cloud deployments.

Talk to an Arm expert
Specifications

Arm AGI CPU specifications and product brief

AI Summary

Specs Arm AGI CPU 136C (max core count) Arm AGI CPU 128C (TCO optimized) Arm AGI CPU 64C (max mem/core)

SKU

  • SP113012
  • SP113012S
  • SP113012A

Processing cores

  • 136 Neoverse V3
  • 2x 128 SVE
  • 2MB/core L2
  • 128 Neoverse V3
  • 2x 128 SVE
  • 2MB/core L2
  • 64 Neoverse V3
  • 2x 128 SVE
  • 2MB/core L2

CPU architecture

  • Armv9.2
  • bfloat16 and INT8 AI instructions
  • Armv9.2
  • bfloat16 and INT8 AI instructions
  • Armv9.2
  • bfloat16 and INT8 AI instructions

System-level cache

  • 128MB
  • 128MB
  • 128MB

Max Frequency

  • 3.5GHz
  • 3.5GHz
  • 3.7GHz

Base TDP*

  • 300W
  • 300W
  • 300W

RDIMM memory

  • 12x DDR5
  • Up to 8800 MT/s
  • 12x DDR5
  • Up to 8800 MT/s
  • 12x DDR5
  • Up to 8800 MT/s

Memory Throughput/core

  • 6GB/s per core
  • 6.3GB/s per core
  • 13GB/s per core

PCIe/IO

  • 96x lanes PCIe Gen6
  • CXL 3.0 Type 3
  • 96x lanes PCIe Gen6
  • CXL 3.0 Type 3
  • 96x lanes PCIe Gen6
  • CXL 3.0 Type 3

PCIe control lanes

  • 6x 1 Gen4
  • 6x 1 Gen4
  • 6x 1 Gen4

2-Socket support

  • Yes
  • Yes
  • Yes

2 DIMMS per channel

  • Yes
  • Yes
  • Yes

*Represents a preset TDP value within the configurable TDP range

Download product brief
Talk to an expert

Talk to an Arm expert to explore how the Arm AGI CPU
is built for next-generation AI data centers.

Contact us

Key takeaways

Key takeaways

  • Purpose-built for agentic AI: The first production silicon from Arm is designed for continuous, large-scale AI systems where CPUs orchestrate thousands of parallel tasks across distributed infrastructure.

  • Rack-scale performance and efficiency: Delivers high, sustained performance across dense deployments, with more than 2x performance per rack compared to x86 systems.

  • High-performance cores and memory architecture: Up to 136 Neoverse V3 cores with class-leading 6GB/s memory bandwidth per core at sub-100ns latency.

  • Designed for AI data center efficiency: Operates at 300W TDP, balancing compute density, memory throughput, and system-level efficiency.

FAQ

Frequently asked questions

Q: What is the Arm AGI CPU and what makes it different from traditional data center CPUs?

A: The Arm AGI CPU is Arm’s first production silicon, designed specifically for agentic AI workloads, delivering high performance, scalable parallel processing, and energy-efficient operation. It enables data centers to run continuous AI workloads at scale while optimizing throughput, resource utilization, and power consumption.

Q: What are the key features of the Arm AGI CPU?

A: The Arm AGI CPU combines high core density, optimized memory architecture, and scalable system design to support AI workloads at scale:

  • Efficient cores: Up to 136 Arm Neoverse V3 cores with dedicated 2 MB L2 cache per core and up to 3.7GHz boost frequency, enabling responsive, parallel performance—Arm AGI CPU specs (source required).
  • Performance and efficiency: High instruction-per-cycle execution on a TSMC 3 nm process with 300W TDP, balancing compute throughput and energy efficiency—Arm AGI CPU specs (source required).
  • Tuned memory architecture: Delivers 6GB/s memory bandwidth per core with support for DDR5-8800 memory and sub-100ns latency, reducing data bottlenecks for AI workloads—Arm AGI CPU specs (source required).
  • Built for density: Supports high-density deployments, including reference designs with up to 272 dedicated cores per 1U server and air-cooled configurations compatible with standard infrastructure—Arm AGI CPU specs (source required).
  • Rack-scale deployment: Designed for large-scale AI infrastructure, enabling configurations with thousands of cores per rack for continuous AI processing—Arm AGI CPU specs (source required).
  • Flexible I/O: Includes 96 PCIe Gen6 lanes, CXL 3.0 for memory expansion, and AMBA CHI links for accelerator connectivity, enabling composable AI systems—Arm AGI CPU specs (source required).

Q: What is agentic AI, and why does it require a new type of CPU?

A: Agentic AI refers to systems that operate continuously, coordinating tasks and making decisions in real time. These workloads require CPUs that can orchestrate distributed systems efficiently at scale.

Q: How does the Arm AGI CPU improve data center performance?

A: It improves performance by delivering high, per-task efficiency and scaling across thousands of cores within a rack, enabling more work per system. This results in more than 2x performance per rack compared to x86 systems.

Q: How does the Arm AGI CPU support AI infrastructure at scale?

A: It supports AI infrastructure by managing distributed workloads, coordinating accelerators, and optimizing data movement across systems to enable continuous, large-scale AI operations.

Q: How does the Arm AGI CPU fit into the Arm ecosystem?

A: It extends the Arm compute platform into production silicon, giving partners the flexibility to deploy Arm technology through IP, compute subsystems, or ready-to-deploy CPUs.


  1. Based on estimates.