FREE SHIPPING on Orders Over US$99

English English

NVIDIA MQM8790-HS2R Mellanox Quantum™ HDR InfiniBand Switch, 40 QSFP56 Ports, 2 Power Supplies (AC), Unmanaged, Standard Depth, C2P Airflow, Rail Kit
P/N: MQM8790-HS2R
US$3899.00
3
Reviews
2
Questions
safeguard
4-Year Warranty
30-Day Returns
30-Day Exchange
(10 In Stock)

Product Highlights

40 HDR 200Gb/s Ports Split into 80×100Gb/s Via Certified Breakout Cables
16Tb/s Non-blocking Switching Fabric with sub-130ns Ultra-low Latency
C2P Airflow Design with Hot-swappable Dual Power Modules
Backward Compatibility with EDR/HDR100 Via Adaptive Speed Negotiation
Unmanaged Architecture for Plug-And-Play Deployment in Automated Clusters
NVIDIA MQM8790-HS2R Mellanox Quantum™ HDR InfiniBand Switch, 40 QSFP56 Ports, 2 Power Supplies (AC), Unmanaged, Standard Depth, C2P Airflow, Rail Kit
40-port non-blocking managed HDR 200Gb/s InfiniBand smart switch, Mellanox provides the world’s smartest switch, enabling in-network computing through the Co-Design Scalable Hierarchical Aggregation and Reduction Protocol (SHARP)™ technology. QM8790 has the highest fabric performance available in the market with up to 16Tb/s of non-blocking bandwidth with sub-130ns port-to-port latency.
Specifications
Manufacturer
Mellanox
Ports
40×QSFP56
Switching Capacity
16Tb/s
Fan Number
5+1 Hot-swappable
Input Voltage
100-127VAC 50/60Hz 4.5A, 200-240VAC 50/60Hz 2.9A
Software
MLNX-OS
Data Rate
40/56/100/200 Gb/s
Latency
130ns
Management
Unsupport
Airflow
Front-to-Back (C2P)
AC Power Supplies
1+1 Hot-swappable
Operating Temperature
0 to 40°C (32 to 104°F)
Storage Temperature
-40 to 70°C (-40 to 158°F)
Dimensions (H×W×D)
43.6×433.2×590.6mm
Rack Units
1 RU
Certification
Product
Product
Product
Product
Product
Product
Product
Product
Product
Product
Questions & Answers
edit icon Ask a question
All (2) Management (1)Product (1)
Q:

What is the maximum bandwidth supported by NVIDIA InfiniBand HDR switches?

by r***m on 2025-07-21 16:32:54

A:

NVIDIA HDR InfiniBand switches support up to 200Gb/s per port, delivering extremely high throughput for data-intensive workloads.

by CPLIGHT on 2025-07-21 16:49:20

like icon Helpful 0
comment icon Comment 0
Q:

Are these switches suitable for large-scale AI and HPC clusters?

by Z***m on 2025-07-21 15:56:30

A:

Yes. NVIDIA InfiniBand switches are designed for high-performance computing and AI workloads, offering low latency and high bandwidth essential for multi-GPU or multi-node training.

by Helpful on 2025-07-21 16:49:43

like icon Helpful 0
comment icon Comment 0
Customer Reviews
edit icon Write a review
All (3) Performance (1)Compatibility (1)Scalability (1)
G
G***m
2025-06-30 06:05:29
Confirmed Purchase
5.0

We were able to seamlessly scale our InfiniBand fabric using NVIDIA’s modular switch design. Adding more compute nodes was straightforward and non-disruptive.

link icon Helpful 0
comment icon Comment 0
i
i***m
2025-05-21 21:41:06
Confirmed Purchase
5.0

Interoperates flawlessly with our NVIDIA NICs and DGX systems. It was truly plug-and-play with minimal configuration needed.

link icon Helpful 0
comment icon Comment 0
O
O***m
2025-04-13 03:08:15
Confirmed Purchase
5.0

The low-latency and high-throughput performance of the NVIDIA InfiniBand switches exceeded expectations. We achieved full HDR bandwidth with consistent sub-microsecond latency.

link icon Helpful 0
comment icon Comment 0