From c2829a54210f1af24ce9c12d106e343b401e9614 Mon Sep 17 00:00:00 2001 From: Csaba Kecskemeti Date: Fri, 23 Jan 2026 19:05:48 -0800 Subject: [PATCH] playbook rev3 --- nvidia/heterogeneous-distributed-inference-rdma/README.md | 3 +++ 1 file changed, 3 insertions(+) diff --git a/nvidia/heterogeneous-distributed-inference-rdma/README.md b/nvidia/heterogeneous-distributed-inference-rdma/README.md index ea86d4c..44bfb8c 100644 --- a/nvidia/heterogeneous-distributed-inference-rdma/README.md +++ b/nvidia/heterogeneous-distributed-inference-rdma/README.md @@ -61,6 +61,9 @@ GPU memory → PCIe → NIC (mlx5) → wire → NIC → PCIe → GPU memory - One QSFP cable (QSFP56 ↔ QSFP28 compatible, 100 Gbps negotiated) - Direct connection or dedicated switch +> [!NOTE] +> **About the hardware used in this tutorial:** We used a ConnectX-5 (MCX516A-CDAT, 100GbE dual-port) on the workstation because that's what we had available. This limits the link speed to 100 Gbps. If you use a ConnectX-7 NIC on the workstation side (matching the DGX Spark), you can achieve up to 200 Gbps. The setup process is the same or very similar - just with higher bandwidth. + > [!NOTE] > Interface names (e.g., `enp1s0f0np0`, `rocep1s0f0`) are system-specific and will differ on your hardware. Use these commands to identify your interfaces: > ```bash