Cisco C885 M8 rack server

Dense AI GPU servers

Power massive GPU density on NVIDIA HGX and AMD OAM platforms

Train trillion-parameter LLMs, run advanced simulations, and more with dense AI GPU servers that deliver interconnect speed and efficiency for even the most ambitious AI workloads.

Purpose-built for the pinnacle of AI performance


01:50
dense ai gpu servers

Dense AI GPU servers with NVIDIA HGX and AMD OAM technology

Engineered to be the heart of large-scale AI infrastructure, Cisco AI GPU servers enable training, fine-tuning, and deployment of massive models fast, efficiently, and at unprecedented scale.

Largest-model optimization

Get exceptional GPU parallelism and bandwidth for cutting-edge use cases, from LLMs, deep learning, and generative AI to scientific simulations.

Architectural choice

Select NVIDIA HGX for NVLink-connected GPU meshes or choose AMD OAM MI300/MI350X accelerators for modular GPU density—both fully integrated into enterprise-ready Cisco platforms.

Rack-scale efficiency

Maximize performance per square foot of data center space with server designs that carefully balance power, cooling, and form factor.

Unified operations

Cisco dense AI GPU servers are integrated with Cisco Intersight for cloud-based lifecycle management, monitoring, and optimization across your entire AI cluster.

Choose your dense AI GPU platform

Cisco dense AI GPU servers support two leading accelerator frameworks: NVIDIA HGX for ultra-high bandwidth GPU meshes and AMD OAM MI300/MI350X for flexible AMD GPU clusters. Both are designed for tight integration, performance, and enterprise-grade management.

  1. NVIDIA HGX – GPU mesh excellence: NVLink-enabled HGX modules deliver unmatched inter-GPU throughput—ideal for training large-scale models with full-bandwidth interconnect.
  2. AMD OAM MI300/MI350X – Modular GPU compute: Open Accelerator Modules from AMD give you modular, interchangeable GPU units with dedicated NVLink-equivalent bandwidth—optimized for dense AI workloads on AMD.
  3. Cisco integration – Ready for the rack: Engineered with power, cooling, and automated management in mind—ensuring seamless deployment from bare metal to full AI rack pod.

Make supercomputing easier with Cisco AI PODs

Cisco AI PODs deliver a validated, modular approach to deploying AI infrastructure quickly and at scale. Available as edge, retrieval-augmented generation (RAG), and large-scale inferencing solutions, AI PODs lower total cost of ownership, provide greater control, and accelerate time to value.

 

Rendering of the VAST Data Platform, multi-tenant, zero-trust, global scale. And VAST DataSpace, edge to cloud namespace. As data from DataStore and DataBase flow into the intelligence in the DataEngine.

Unlock AI insights with the VAST Data Platform

The VAST Data Platform delivers a unified, high-performance data foundation for AI, analytics, and deep learning. Integrated into Cisco AI architectures, it enables massive data throughput, simplifies access to enterprise data, and fuels faster model training and RAG pipelines.


Cisco AI software and solutions

Cisco Intersight

Cloud-operated and on-premises infrastructure management simplifying AI server deployment, monitoring, and optimization.

Cisco Validated Designs for AI

Tested architectures for building secure, high-performance AI infrastructure from edge to cloud faster and with reduced risk.

MLPerf 5.0 benchmarks

Top-notch performance for LLM training

Cisco UCS C885A M8, built on the NVIDIA HGX platform, shines in machine learning performance benchmarks, achieving top rankings for Llama-3.1-405B (H200) and Stable Diffusion XL (H100), with up to 40% faster inference.