News

/
News

Arm AGI CPU: The Chip Foundation of Intelligent Agent-Based Artificial Intelligence in the Cloud Era

Today, Arm officially launched the Arm AGI CPU - this is a brand-new mass-produced chip built on the Arm Neoverse platform, specifically designed to 

empower the next generation of artificial intelligence infrastructure. 

During the over 35-year development history of Arm, this is the first time we have launched our own self-developed chip product. This move will further 

expand the capabilities of the Arm Neoverse platform from IP and Arm computing subsystems (CSS), enabling customers to deploy it. 

The Arm computing architecture offers more options: customers can either independently develop customized chips, integrate platform-level solutions, 

or directly deploy processors designed by Arm. This layout aligns with the requirements of the artificial intelligence infrastructure. 

This rapid evolution trend also meets the growing demand of ecosystem partners for an Arm production-grade platform that can be rapidly scaled up 

for deployment. 

The rise of intelligent agent-based artificial intelligence infrastructure 

Nowadays, artificial intelligence systems are gradually achieving 24/7 continuous operation worldwide. In the past, humans were the bottleneck in the 

computing process - the efficiency of human-computer interaction determined the processing speed of the business within the system. And


In the era of intelligent agent-based artificial intelligence, this limitation has been completely broken: software intelligent agents can autonomously 

coordinate tasks, connect with multiple artificial intelligence models, and make real-time decisions. 

As the continuous operation of artificial intelligence systems and the increasing complexity of workloads continue, the CPU has become the core 

scheduling unit of modern infrastructure, bearing the heavy responsibility of ensuring the efficient and large-scale operation of distributed artificial 

intelligence systems. In


In modern artificial intelligence data centers, the CPU needs to manage thousands of distributed tasks, including coordinating accelerators, controlling 

memory and storage, scheduling workloads, and facilitating data flow across systems; in the context of agent-based artificial intelligence scenarios, 

The CPU also needs to coordinate the task distribution work for a large number of intelligent agents. 

This transformation imposed new requirements on the CPU and also spurred the iterative upgrading of processor technology. 

The Arm Neoverse architecture has become the technical foundation for many of the current top-level ultra-large-scale cloud computing and artificial 

intelligence platforms. Amazon Web Services' Graviton, Google's Axion, Microsoft's Azure Cobalt, and NVIDIA's Vera are all examples. 

Based on this architecture. As the artificial intelligence infrastructure expands globally on a large scale, ecosystem partners are all looking forward to 

Arm further leveraging its technological advantages. And the Arm AGI CPU is precisely designed to address this industry transformation. 

Arm AGI CPU: Designed for intelligent energy efficiency at the cabinet level 


The workload of intelligent agent-style artificial intelligence requires chips to maintain continuous high-performance output at an extremely large scale. 

The design intention of Arm AGI CPU is to enable thousands of computing cores to operate under the power and heat dissipation constraints of modern data centers. 

When running in parallel mode, it can still maintain excellent single-task performance under continuous load. 

From the operating frequency to the memory and input/output architecture, every design detail of the Arm AGI CPU has been meticulously crafted to 

meet the requirements of large-scale parallel and high-performance intelligent agent-style artificial intelligence workloads in high-density cabinet 

deployment scenarios. 


Core technological advantages 


✅ Intelligent Response Performance: Up to 136-core Arm Neoverse V3 processor, each core equipped with 2MB dedicated secondary cache, with a 

maximum clock speed of up to 3.7GHz 

✅ Input/output capabilities for modular artificial intelligence systems: 96 PCIe 6.0 channels, native support for CXL 3.0 (enabling functions such as 

memory expansion), equipped with AMBA CHI expansion links 

✅ Industry-leading Arm energy efficiency: Achieving unparalleled energy efficiency through advanced 3-nanometer manufacturing process, maximizing 

computing density, with a thermal design power of only 300 watts. 

✅ Dual-core design: The memory and input/output modules are integrated onto the same chip. The memory latency is less than 100 nanoseconds, 

and the memory access has been optimized for latency. 

✅ Exceptionally high memory performance: Single-core memory bandwidth is 6GB/s, the maximum memory capacity of a single chip is 6TB, supports 

DDR5-8800 specification, and the memory architecture is specifically optimized for computing scenarios. 


The reference server launched by Arm adopts a 1OU dual-node design. Each blade server integrates two Arm AGI CPUs, and is equipped with dedicated 

memory and input/output modules. The total number of cores reaches 272. This blade server can fully utilize its capabilities. 

The deployment is carried out in a standard 36-kilowatt air-cooled cabinet. With 30 blade servers, a total computing power of 8,160 cores can be achieved. 

Additionally, Arm has collaborated with Supermicro to develop a 200-kilowatt liquid-cooled cabinet solution, which can accommodate 336 chips. 

Arm AGI CPU has a total of over 45,000 cores. 

Under the aforementioned configuration, the single-cabinet performance of the Arm AGI CPU can exceed that of the latest x86 system by more than 

twice. This advantage stems from the inherent technical characteristics of the Arm architecture, as well as the precise matching of system resources with 

computing demands. Specifically 

It is reflected in three aspects: 

The Arm AGI CPU boasts industry-leading memory bandwidth, enabling a single cabinet to support a greater number of effective execution threads; 

while x86 CPUs, under continuous load, experience resource competition among cores, which leads to performance degradation. 

The single-core of the Arm Neoverse V3 processor combines high performance with high energy efficiency, performing better than traditional 

architectures. Each Arm thread can handle more tasks. 

More available threads, combined with the enhanced task processing capability of a single thread, have jointly contributed to a significant improvement 

in the performance of a single cabinet. 

The early positive feedback of the artificial intelligence ecosystem 

The Arm AGI CPU has received high recognition from the leading partners in the artificial intelligence ecosystem, and its commercial implementation 

progress is remarkable. The planning and deployment scenarios of the partners cover accelerator management, intelligent agent task orchestration, and 

the expansion of intelligent agent tasks. 

High-density services / applications / tools, as well as an enhanced network and data plane computing capability specifically designed for artificial 

intelligence data centers. 


**The Meta Universe** (Meta) is our primary partner and initial client. The two parties jointly developed the Arm AGI CPU, aiming to optimize the 

performance for the gigawatt-level infrastructure of Meta's entire range of applications, and to integrate it with Meta's own research and development. 

The MTIA accelerator is operating in a collaborative manner. Other initial partners include Cerebras, Cloudflare, F5, OpenAI, Positron, Rebellions, and SIA. 

PwC (SAP) and SK Telecom of South Korea. These companies have all engaged in in-depth cooperation with Arm, by deploying Arm AGI CPUs, to 

accelerate the implementation of AI-driven services in cloud, network and enterprise environments. Currently, Huoqing 

The companies of Erqi, Lianxi and Chaohui have initiated the commercial system ordering based on the Arm AGI CPU. 


To further accelerate the implementation of the product, Arm has simultaneously launched the Arm AGI CPU 1OU dual-node reference server. This 

product adopts the Open Compute Project (OCP) DC-MHS standard form. Arm plans to use this reference server 

The design plan and associated firmware will be made available for open-source contribution. Additionally, the system architecture specifications, 

debugging framework, as well as diagnostic and verification tools applicable to all Arm-based systems will also be made available. More details will 

be announced at the upcoming OCP conference in Europe. It was announced at the summit of Asia, the Middle East and Africa. 

A brand-new chapter of Arm's infrastructure business 


The release of Arm AGI CPU marks a new stage of development for Arm's data center business, and also demonstrates Arm's continuous leading position 

in the field of computing innovation. In the current era where artificial intelligence is reshaping the entire industry, Arm has always... 

Committed to promoting technological progress across the entire ecosystem, we provide tailored solutions to different types of clients, including 

large-scale cloud service providers and AI startups. 


Arm AGI CPU is the first product of Arm's new data center chip product line. It is now available for order. Arm has already planned the subsequent 

product development roadmap and will continue to strive for industry-leading performance, scalability and energy efficiency. 

At the same time, this product line will be advanced in parallel with the Arm Neoverse CSS product roadmap, ensuring that all Arm data center 

customers can achieve synchronous development in terms of platform architecture and software compatibility. 


Entering a new stage of development, Arm's mission remains unchanged: to build the foundation of computing power and empower the innovation 

of various industries. And this vision of Arm has also received strong support from the entire ecosystem: over 50 companies from large-scale computing, 

The leading enterprises in the fields of cloud computing, chips, memory, networks, software, system design and manufacturing are helping Arm's 

computing platform expand from IP to chip products. With the Arm AGI CPU, we not only define the native artificial intelligence... 

The architecture of the intelligent data center is not only a result of designing this brand-new infrastructure by ourselves. 


Partner's Message: The Practical Value of Arm AGI CPU 


Cerebras 

Syrup focuses on building ultra-high-speed and large-scale artificial intelligence infrastructure for reasoning, which is becoming the mainstream 

workload in the field of artificial intelligence. In this context, modular high-performance systems have become increasingly important - such systems 

It is necessary to have both specialized acceleration chips specifically designed for artificial intelligence, as well as efficient and scalable CPUs to scale 

up the coordination of data flow, network communication, and task scheduling. Arm has expanded its computing platform to the level of general 

artificial intelligence infrastructure. 

For both customers in the field of ecology and those involved in the global large-scale deployment of artificial intelligence, this represents a significant 

positive breakthrough. 

-- Andrew Feldman, CEO of Synopsys 


Cloudflare 

The mission of CloudFlare is to contribute to the creation of a better internet. This requires that our infrastructure be able to achieve efficient expansion 

across the global network. The high-performance and high-energy efficiency computing power provided by Arm AGI CPU is specifically designed for 

the next-generation workloads. 

It is precisely tailored to meet our needs. 

-- Stephanie Cohen, Chief Strategy Officer of Cloudfire 


Metaverse 

To provide an artificial intelligence experience globally, a powerful and flexible customized chip solution portfolio is required. These chips need to be 

specifically tailored to accelerate artificial intelligence workloads and optimize Meta's platform-wide performance. Jointly developing Arm AGI CPU 

with Arm, aiming to create an efficient computing platform, significantly enhancing the performance density of data centers, and providing cross-

generation technical support for our continuously evolving artificial intelligence systems. 

-- Santosh Janardhanan, Head of Meta's Infrastructure Department 


OpenAI

OpenAI operates large-scale artificial intelligence systems, with hundreds of millions of users using ChatGPT every day. Enterprises leverage our API 

for development, and developers rely on tools like Codex for innovation. As the business expands on a larger scale, 

The Arm AGI CPU will become an important component of our infrastructure, enhancing the orchestration layer capability for coordinating large-scale 

artificial intelligence workloads, and significantly improving the system's energy efficiency, performance and bandwidth. 

-- Sachin Katyi, Head of Industrial Computing Department at OpenAI 


Positron 

Positrons are dedicated to developing customized inference accelerators that can achieve breakthrough token generation efficiency through general-

purpose memory. Arm has always been able to launch the most energy-efficient computing platforms in the industry, which makes the Arm AGI CPU 

the next-generation human-like 

The natural foundation of artificial intelligence infrastructure. By combining positron's reasoning acceleration technology with the high-efficiency Arm 

AGI CPU platform, we expect to help data center operators achieve better performance per watt and per dollar. 

Carry out large-scale deployment of cutting-edge artificial intelligence models. 

-- Mitesh Agrawal, Chief Executive Officer of Positron AI 


Rebellions Control Technology 

High-performance artificial intelligence systems require close collaboration between general-purpose computing and accelerator architectures. By 

integrating the Arm AGI CPU with the Renwei Intelligent Control's neural network processor (NPU) into a brand-new high-density server configuration, 

we are Create a scalable and energy-efficient platform that has been specifically optimized for large-scale artificial intelligence inference workloads. 

-- Marshall Cai, Chief Commercial Officer of RuiLink Intelligent Control 


SAP 

SAP has successfully deployed SAP HANA on the Amazon Web Services Graviton platform based on the Arm architecture. This fully demonstrates the 

maturity and performance of the Arm ecosystem in enterprise-level workload scenarios. 

-- Stefan Baller, Senior Vice President and Head of the HANA and Persistence Technology Department at SAP 


SK Telecom of South Korea 

SK Telecom is building a large-scale, full-stack artificial intelligence inference data center infrastructure, which includes Arm AGI CPUs and Ruilian 

Zhikong artificial intelligence acceleration chips. We will also independently develop the A.X large model and integrate it with the optimized version. 

Combining the optimized artificial intelligence server with this solution is expected to not only bring it to the global market, but also significantly 

enhance the competitiveness of our artificial intelligence data center in the industry. 

-- Zheng Xigen, Chief Technology Officer of SK Telecom and Head of the Artificial Intelligence Innovation Center