Thursday, April 10th 2025

AMD Pensando Pollara 400 AI NIC Now Available and Shipping to Customers

To effectively train and deploy generative AI, large language models, or agentic AI, it's crucial to build parallel computing infrastructure that offers the best performance to meet the demands of AI/ML workloads but also offers the kind of flexibility that the future of AI demands. A key aspect for consideration is the ability to scale-out the intra-node GPU-GPU communication network in the data center.

At AMD, we believe in preserving customer choice by providing customers with easily scalable solutions that work across an open ecosystem, reducing total cost of ownership—without sacrificing performance. Remaining true to that ethos, last October, we announced the upcoming release of the new AMD Pensando Pollara 400 AI NIC. Today we're excited to share the industry's first fully programmable AI NIC designed with developing Ultra Ethernet Consortium (UEC) standards and features is available for purchase now. So, how has the Pensando Pollara 400 AI NIC been uniquely designed to accelerate AI workloads at scale?
Building High-Performing AI Infrastructure
Cloud service providers, hyperscalers, and enterprises are looking to maximize the performance from their AI clusters. However, the network has been cited by many as a primary bottleneck for GPU utilization issues. Data transfer speed only matters if the network is properly optimized to capitalize on it.

As AI workloads continue to grow at an astounding rate, organizations cannot afford to underutilize networking and compute resources. The three main attributes of networks with the highest rates of utilization are delivering on intelligent load balancing, congestion management, fast failover and loss recovery. Highly performant networks must also be continually optimized for increased uptime, job completion times, reliability, availability and serviceability—at scale.

Extensible Future-Ready Infrastructure
Hardware Programmability Powering Customer Roadmaps
Powered by our P4 architecture, the Pensando Pollara 400 AI NIC offers a fully programmable hardware pipeline. This powers the ultimate customer flexibility-from adding new capabilities, like those issued by the UEC and/or developing custom transport protocols, designed to accelerate company roadmaps. Now as new standards, company initiatives, or AI workloads emerge, customers don't have to wait for the next generation AI NIC hardware to accelerate their workloads.

Developing UEC Features for Accelerating Next-Gen Workloads
  • Transport Protocol of Choice: tap into RoCEv2, UEC RDMA, or any Ethernet protocol of your choice.
  • Intelligent Packet Spray: a feature to increase network bandwidth utilization with advanced adaptive packet spraying, which is critical for managing the high bandwidth and low latency required by large AI models.
  • Out-of-Order Packet Handling and In-order Message Delivery: designed to reduce buffer time by intelligently managing out-of-order packet arrivals, a common challenge associated with multipathing and packet spraying techniques, this feature minimizes errors and enhances efficiency during AI training and inference, all without relying on scale-out switching fabric.
  • Selective Retransmission: enhance network performance by only resending lost or corrupted packets through in-order message delivery and selective acknowledgment (SACK) retransmission.
  • Path-Aware Congestion Control: optimize network performance with intelligent load balancing automatically avoiding congested paths and help sustain near wire-rate performance during transient congestion.
  • Rapid Fault Detection: accelerate AI job completion times by issues in milliseconds, with sender-based ACK monitoring, receiver-based packet monitoring, and probe-based verification, enabling near-instantaneous failover, minimizing GPU idle time.
The Open Ecosystem Advantage
By offering vendor agnostic compatibility, organizations can build AI infrastructure that meets the demands for current workloads and offers easy scalability and programmability for future requirements. With this open ecosystem approach, the AMD AI NIC helps reduce CapEx without sacrificing performance and without a dependency to deploy expensive cell based, large buffer switching fabrics.

Validated in Some of The Largest Scale-out Data Centers
Last, but certainly not least—the Pensando Pollara 400 AI NIC is set to power some of the largest scale-out infrastructure with the first customer shipments having gone to and been tested by some of the largest Cloud Service Providers (CSPs) in the world. CSPs chose the Pensando Pollara 400 AI NIC due to its distinctive programmability, high-bandwidth, low latency performance, rich feature set, and truly extensible infrastructure compatible across an open ecosystem.
Sources: AMD Blog, AMD, AMD Twitter
Add your own comment

14 Comments on AMD Pensando Pollara 400 AI NIC Now Available and Shipping to Customers

#1
jak_2456
Pensando means "thinking" in Spanish.
Posted on Reply
#2
ThomasK
jak_2456Pensando means "thinking" in Spanish.
As well as Portuguese.
Posted on Reply
#3
AusWolf
I hope products like this will move AI people away from gaming GPUs.
Posted on Reply
#4
Visible Noise
AusWolfI hope products like this will move AI people away from gaming GPUs.
This is a networking card, it has nothing to do with AI compute.

NIC = Network Interface Card.
Posted on Reply
#5
igormp
I wonder how flexible this product is when compared to Nvidia's BlueField.
Given that this is using xilinx's IP, I wonder if it features some FPGA-esque stuff as well, but I never had any experience with pensando products.
Posted on Reply
#6
AnarchoPrimitiv
Can somebody explain to me like I'm a child, how this NIC differs from a typical NIC
Posted on Reply
#7
Wirko
AnarchoPrimitivCan somebody explain to me like I'm a child, how this NIC differs from a typical NIC
It's a typical NIC, and its price (per Gbit/s) is probably similar to common NICs working at 10 Gbit/s.
Posted on Reply
#8
R-T-B
AnarchoPrimitivCan somebody explain to me like I'm a child, how this NIC differs from a typical NIC
It's faster (400Gbits) to facilitate AI across multiple machines.
Posted on Reply
#9
Dr. Dro
igormpI wonder how flexible this product is when compared to Nvidia's BlueField.
Given that this is using xilinx's IP, I wonder if it features some FPGA-esque stuff as well, but I never had any experience with pensando products.
There's a product brief pdf on AMD's site. Uses a QSFP-112 interface, but it doesn't seem to have the large amount of onboard memory that BlueField does. It's also slower than BlueField-4.

www.amd.com/content/dam/amd/en/documents/pensando-technical-docs/product-briefs/pensando-pollara-400-product-brief.pdf

It's not Xilinx IP, btw. It's from the "startup" (well, comparably small and new) Pensando Systems, which AMD acquired in an all-cash transaction shortly after they acquired Xilinx.

www.datacenterdynamics.com/en/news/amds-19bn-acquisition-of-data-center-smartnic-firm-pensando-closes/
AnarchoPrimitivCan somebody explain to me like I'm a child, how this NIC differs from a typical NIC
Ultra high speed, GPU to GPU hyperscaling interconnect is the primary use, these are "DPUs", some form of hyper fast Ethernet card on steroids
Posted on Reply
#10
AusWolf
Visible NoiseThis is a networking card, it has nothing to do with AI compute.

NIC = Network Interface Card.
Fair point, thanks.

Maybe if these walls of marketing text were a bit shorter, or written by someone working for TPU, I would actually bother reading them.
Posted on Reply
#11
Jermelescu
jak_2456Pensando means "thinking" in Spanish.
You mean Mexican?
ThomasKAs well as Portuguese.
And Brazilian?
Posted on Reply
#12
Wirko
AusWolfFair point, thanks.

Maybe if these walls of marketing text were a bit shorter, or written by someone working for TPU, I would actually bother reading them.
If you know how to use an AI debuzzworder, there's a lot of useful info hidden in here:

community.amd.com/t5/corporate/transforming-ai-networks-with-amd-pensando-pollara-400/ba-p/716566

In my understanding, this card *at least* handles the protocol(s) at ISO/OSI Layer 4 (Transport layer),where TCP resides in common Ethernet. Leaving that job to CPU cores might be too inefficient at the speed of 50 GB/s, with too high average latencies or maximum latencies. It probably handles some higher-level networking layer(s) too, might also help to achieve advanced things like memory coherence across nodes, and does "congestion control", which is what routers usually do.
Posted on Reply
#13
AusWolf
WirkoIf you know how to use an AI debuzzworder, there's a lot of useful info hidden in here:

community.amd.com/t5/corporate/transforming-ai-networks-with-amd-pensando-pollara-400/ba-p/716566

In my understanding, this card *at least* handles the protocol(s) at ISO/OSI Layer 4 (Transport layer),where TCP resides in common Ethernet. Leaving that job to CPU cores might be too inefficient at the speed of 50 GB/s, with too high average latencies or maximum latencies. It probably handles some higher-level networking layer(s) too, might also help to achieve advanced things like memory coherence across nodes, and does "congestion control", which is what routers usually do.
AI debuzzworder... That's the first useful AI tool I've heard about. It's a shame we need one these days.
Posted on Reply
#14
Operandi
AnarchoPrimitivCan somebody explain to me like I'm a child, how this NIC differs from a typical NIC
Its basically a NIC that takes a bunch of the workload off of the CPU, the term is DPU. Serve The Home can explain it better than I can but its going to be a lot more common in datacenters depending on the workload.
Posted on Reply
Add your own comment
May 5th, 2025 08:51 EDT change timezone

New Forum Posts

Popular Reviews

Controversial News Posts