Mellanox (NVIDIA Mellanox) 920-9B210-00FN-0D0 InfiniBand Switch in Production

April 15, 2026

Laatste bedrijfsnieuws over Mellanox (NVIDIA Mellanox) 920-9B210-00FN-0D0 InfiniBand Switch in Production
Background & Challenge: The Scalability Wall in AI Training Clusters

A leading East Asian AI research institute faced a common but critical bottleneck. Their 512-GPU cluster, used for large language model training and molecular dynamics simulations, was suffering from severe performance degradation as jobs scaled. The root cause was the legacy 100Gb/s Ethernet fabric, where TCP/IP overhead and packet loss during incast events caused GPU idle times of up to 35%. The team needed a lossless, ultra-low-latency fabric that could support RDMA and scale to thousands of nodes without compromising on deterministic performance. After evaluating several solutions, they selected the Mellanox (NVIDIA Mellanox) 920-9B210-00FN-0D0 InfiniBand switch as the core of their new spine-leaf architecture.

Solution & Deployment: Integrating the 920-9B210-00FN-0D0 into an NDR Fabric

The deployment centered around the 920-9B210-00FN-0D0 as the spine layer, with 32 leaf switches connecting 512 NVIDIA A100 GPUs via ConnectX-7 adapters. Each 920-9B210-00FN-0D0 MQM9790-NS2F 400Gb/s NDR switch provides 400Gb/s per port, doubling the bandwidth of previous HDR solutions while maintaining sub-microsecond switching latency. The official 920-9B210-00FN-0D0 InfiniBand switch OPN simplified procurement and ensured firmware consistency across all units. Network engineers utilized the detailed 920-9B210-00FN-0D0 datasheet and 920-9B210-00FN-0D0 specifications to validate power and thermal requirements, enabling a seamless integration into existing 19" racks. Crucially, the switch is fully 920-9B210-00FN-0D0 compatible with both the existing HDR infrastructure and newer NDR endpoints, allowing for a phased migration.

Performance Results & Operational Benefits
  • RDMA Efficiency Gains: With the NVIDIA Mellanox 920-9B210-00FN-0D0 enabling hardware-based congestion control, RDMA write latency dropped from 12µs to 1.2µs. GPU direct RDMA (GDR) became fully effective, eliminating CPU memory bottlenecks.
  • HPC Application Speedup: A key weather modeling code (MPI-based) saw a 2.7x performance improvement due to the switch's adaptive routing and SHARP v2 collective offloads.
  • AI Training Throughput: For a 175-billion-parameter LLM training job, the new fabric reduced all-reduce time by 68%, improving overall GPU utilization from 62% to 91%.
  • Operational Simplicity: The 920-9B210-00FN-0D0 InfiniBand switch OPN solution integrated with NVIDIA's UFM platform, providing real-time telemetry and predictive failure alerts. IT managers reported a 50% reduction in network-related troubleshooting time.
Cost & Availability Insights

When evaluating the project, the research institute benchmarked the 920-9B210-00FN-0D0 price against competing Ethernet solutions. Despite a higher upfront cost, the total cost of ownership (TCO) favored InfiniBand due to higher GPU utilization and lower power per Gb/s. Units are readily available as 920-9B210-00FN-0D0 for sale through NVIDIA's distribution channels, with lead times significantly shorter than other NDR switches. The detailed 920-9B210-00FN-0D0 specifications also confirmed support for redundant power supplies and hot-swappable fans, meeting the institute's reliability requirements for 24/7 AI research operations.

Key Technical Specifications at a Glance
Parameter Detail
Model NVIDIA Mellanox 920-9B210-00FN-0D0
Port Speed 400Gb/s NDR (per port)
Base OPN 920-9B210-00FN-0D0 InfiniBand switch OPN
Full Config 920-9B210-00FN-0D0 MQM9790-NS2F 400Gb/s NDR
Summary & Outlook: The Standard for Next-Gen AI Fabrics

The AI research institute has now standardized on the 920-9B210-00FN-0D0 for all future cluster expansions, including a planned 2,048-GPU NDR200 fabric. This real-world case demonstrates that the NVIDIA Mellanox 920-9B210-00FN-0D0 is not just a switch — it is a foundational component for achieving linear performance scaling in AI and HPC environments. For architects and IT managers looking to eliminate networking bottlenecks, the 920-9B210-00FN-0D0 InfiniBand switch OPN solution offers a proven, production-ready path forward.