Cisco continues to redefine AI infrastructure with the launch of the UCS C880A M8 Rack Server, now equipped with the ultra-powerful NVIDIA HGX B300 SXM GPUs and the latest Intel Xeon 6th-Generation (Granite Rapids) CPUs. This marks a convergence of state-of-the-art GPU acceleration, high-throughput CPU compute, and enterprise-grade infrastructure management, optimized for modern AI workloads.
NVIDIA: HGX B300 — Unprecedented AI Performance
Drawing on NVIDIA’s messaging around the HGX B300 platform:
- Inference performance: Up to 11× higher throughput for models like Llama 3.1 405B versus the previous Hopper generation, thanks to Blackwell Tensor Cores and Transformer Engine optimizations.
- Training speed: Up to 4× faster for large-scale LLMs like Llama 3.1 405B via FP8 precision enhancements, NVLink 5 interconnects (with 1.8 TB/s bandwidth), InfiniBand, and Magnum IO software.
- SXM form factor advantage: High-bandwidth socket architecture enables superior GPU-to-GPU NVLink connectivity and power delivery without cabling, minimizing bottlenecks and simplifying scaling.
Intel: Xeon 6th-Gen CPUs — CPU Power Meets AI Acceleration
Powered by the newly released Intel Xeon 6 (6700P/6500P) P-core processors, the UCS C880A M8 delivers:
- Up to 86 cores per socket, doubled memory bandwidth, and integrated AI accelerators—including Trust Domain Extensions (TDX), Advanced Matrix Extensions (AMX), Data Streaming Accelerator (DSA), QuickAssist (QAT), and In-Memory Analytics Accelerator (IAA).
- Support for DDR5-6400 and MRDIMMs, boosting memory throughput further.
- Optimized for compute-intensive and hybrid AI workloads—from inference pipelines to large-scale training nodes.
Cisco: Intersight Management + AI POD Integration
Cisco’s AI infrastructure goes beyond raw compute:
- The UCS C880A M8 integrates seamlessly with Cisco Intersight, Cisco’s SaaS-based management platform offering centralized control, visibility, and policy management across the distributed AI stack.
- It fits within Cisco AI PODs, modular AI infrastructure solutions designed for rapid deployment at scale. These validated data center units simplify AI factory rollouts while ensuring interoperability with compute, networking, and security.
Key Use Cases Enabled by HGX B300 (SXM)
The integration of HGX B300 SXM GPUs and Xeon 6th-Gen CPUs enables a suite of AI workloads:
- Real-Time LLM Inference
- Run massive models like Llama 3.1 405B with ultra-low latency and high throughput, ideal for chatbots, agents, and real-time reasoning.
- Large-Scale Model Training & Fine-Tuning
- Take advantage of 4× faster training and massive inter-GPU bandwidth to train or fine-tune models with hundreds of billions of parameters.
- High-Performance AI Pipelines
- Leverage CPU offload for data preparation and orchestration, alongside GPU acceleration, for ETL, multimodal processing, and inference workflows.
- AI-Native Data Centers / AI Factories
- Build composable, secure, and scalable AI infrastructure blocks with Cisco AI PODs, ready for integration in data centers or at the edge.
- HPC & Scientific Simulation
- Run contiguous memory models and multi-GPU workloads with enhanced NVLink connectivity for high fidelity simulations and analytics.
Summary Table
Component |
Highlight |
GPU |
NVIDIA HGX B300 SXM: 11× inference, 4× training, NVLink 5 bandwidth — leading AI acceleration |
CPU |
Intel Xeon 6th-Gen P-core (up to 86 cores), DDR5-6400, built-in AI accelerators |
Platform |
Cisco UCS C880A M8 with Intersight integration—scalable, orchestrated, and enterprise-ready |
Ecosystem |
Cisco AI PODs + Secure AI Factory + strong interconnect (network, security, validation) |
Use Cases |
LLM inference/training, AI pipelines, AI POD deployment, HPC workloads |
Final Thoughts
The Cisco UCS C880A M8: HGX B300 with Intel Xeon 6th-Gen sets a new benchmark in AI infrastructure. It offers hyperscale-level AI performance, rock-solid CPU support, enterprise-grade manageability via Intersight, and secure deployments through Cisco Secure AI factory with Nvidia and Cisco AI PODs scalable architectures. Whether you’re building an AI training cluster, LLM inference engine, or composable AI infrastructure, this platform is purpose-built for the next frontier of AI.
Discover the power of next-gen AI infrastructure—read the Cisco UCS C880A M8 Data Sheet
We’d love to hear what you think. Ask a Question, Comment Below, and Stay Connected with #CiscoPartners on social!
Cisco Partners Facebook | @CiscoPartners X | Cisco Partners LinkedIn