CHAPTER 1
Overview
This chapter contains the following topics:
• Overview, on page 1
• External Features, on page 7
• Component Locations, on page 13
• Summary of Features, on page 28
Overview
The Cisco UCS C885A M8 Rack Server is a dense-GPU server designed to deliver massive, scalable accelerated
compute capabilities to address the most demanding AI workloads, including deep learning/Large Language
Model (LLM) training, model fine-tuning, large model inferencing, and Retrieval-Augmented Generation
(RAG).
To deliver massive accelerated compute performance in a single server, the server offers a choice of eight
GPUs of the following types:
• NVIDIA
®
H100 SXM or NVIDIA
®
H200 Server PCI Express Module (SXM) GPUs. SXM is a
socket-based GPU interconnect method used by NVIDIA GPUs.
• AMD MI300X OCP Accelerator Model (OAM) GPUs. OAM is an Open Compute GPU interconnect
standard that avoids GPU vendor lock in.
For north-south traffic, the server supports one NVIDIA Bluefield-3 B3220 DPU. Eight NVIDIA ConnectX-7s
or Bluefield-3 B3140H SuperNICs are supported for east-west traffic between GPUs enabling AI model
training across a cluster of dense-GPU servers.
The server is offered in fixed configurations with resources optimized for intensive AI and HPC workloads.
Cisco UCS C885A M8 Server Installation and Service Guide
1