playbook rev2

This commit is contained in:
Csaba Kecskemeti 2026-01-23 18:59:01 -08:00
parent c5d7b777e1
commit b254e6f632
3 changed files with 16 additions and 10 deletions

View File

@ -384,7 +384,17 @@ python -m vllm.entrypoints.openai.api_server \
--- ---
## Step 8. Run Production Model (72B) ## Step 8. Run Large Model (72B)
This step demonstrates the real power of distributed inference: running a model that **exceeds the memory capacity of any single GPU**.
| Component | Available VRAM | Sufficient for 72B? |
|-----------|---------------|---------------------|
| DGX Spark | 128 GB | No (~136GB needed) |
| RTX 6000 Pro | 96 GB | No (~136GB needed) |
| **Combined Cluster** | **224 GB** | **Yes** |
The Qwen2.5-72B-Instruct model requires ~136GB in BF16 precision - impossible to run on either GPU alone. This is where our RDMA cluster shines, aggregating memory across both systems.
Memory-optimized configuration for 136GB model: Memory-optimized configuration for 136GB model:
@ -524,5 +534,3 @@ This playbook was contributed by **Csaba Kecskemeti** | [DevQuasar](https://devq
For a detailed walkthrough and additional context, see the original article: For a detailed walkthrough and additional context, see the original article:
[Distributed Inference Cluster: DGX Spark + RTX 6000 Pro](https://devquasar.com/ai/edge-ai/distributed-inference-cluster-dgx-spark-rtx-6000-pro/) [Distributed Inference Cluster: DGX Spark + RTX 6000 Pro](https://devquasar.com/ai/edge-ai/distributed-inference-cluster-dgx-spark-rtx-6000-pro/)
![DevQuasar](assets/devquasar-logo.png)

View File

@ -27,8 +27,8 @@ GPU memory → PCIe → NIC (mlx5) → wire → NIC → PCIe → GPU memory
**Key properties:** **Key properties:**
- **No CPU copies:** Data bypasses system memory - **No CPU copies:** Data bypasses system memory
- **No kernel networking stack:** Direct hardware-to-hardware communication - **No kernel networking stack:** Direct hardware-to-hardware communication
- **Ultra-low latency:** ~750 nanoseconds end-to-end - **Ultra-low latency:** Microsecond-level communication
- **High message rate:** Up to 200M messages/second - **High throughput:** 93+ Gbps validated over 100 Gbps link
## What you'll accomplish ## What you'll accomplish
@ -135,7 +135,7 @@ Both planes use the same 100 Gbps ConnectX network in this configuration.
1. Install the ConnectX card in a PCIe Gen3/4 x16 slot (CPU-direct, not via chipset) 1. Install the ConnectX card in a PCIe Gen3/4 x16 slot (CPU-direct, not via chipset)
2. **Cooling Requirements:** ConnectX-5 100GbE cards generate significant heat under load. Ensure adequate case airflow and monitor temperatures with `sensors | grep mlx` 2. **Cooling Requirements:** ConnectX-5/7 100GbE cards are primarily designed for server environments with active cooling. In a workstation, ensure adequate case airflow directed at the card, and consider adding a PCIe slot fan for sustained high-bandwidth workloads.
3. **BIOS settings:** 3. **BIOS settings:**
``` ```
@ -416,7 +416,7 @@ sudo ip link set enp1s0f0np0 up
sudo ip link set enp1s0f0np0 mtu 9000 sudo ip link set enp1s0f0np0 mtu 9000
``` ```
**Option 2: Permanent Configuration (Production)** **Option 2: Permanent Configuration**
First, identify your active internet interface on both systems: First, identify your active internet interface on both systems:
@ -549,7 +549,7 @@ Example successful output:
- Link type: Ethernet confirms RoCE v2 is working - Link type: Ethernet confirms RoCE v2 is working
**Performance expectations:** **Performance expectations:**
- **>90 Gbps:** Excellent - Ready for production AI workloads - **>90 Gbps:** Excellent - Ready for distributed AI workloads
- **80-90 Gbps:** Good - Sufficient for most multi-node training - **80-90 Gbps:** Good - Sufficient for most multi-node training
- **<80 Gbps:** Check MTU (should be 9000), cable quality, or PCIe slot - **<80 Gbps:** Check MTU (should be 9000), cable quality, or PCIe slot
@ -643,5 +643,3 @@ This playbook was contributed by **Csaba Kecskemeti** | [DevQuasar](https://devq
For a detailed walkthrough and additional context, see the original article: For a detailed walkthrough and additional context, see the original article:
[Distributed Inference Cluster: DGX Spark + RTX 6000 Pro](https://devquasar.com/ai/edge-ai/distributed-inference-cluster-dgx-spark-rtx-6000-pro/) [Distributed Inference Cluster: DGX Spark + RTX 6000 Pro](https://devquasar.com/ai/edge-ai/distributed-inference-cluster-dgx-spark-rtx-6000-pro/)
![DevQuasar](assets/devquasar-logo.png)

Binary file not shown.

Before

Width:  |  Height:  |  Size: 35 KiB