diff --git a/nvidia/nccl/README.md b/nvidia/nccl/README.md index a293075..9554130 100644 --- a/nvidia/nccl/README.md +++ b/nvidia/nccl/README.md @@ -70,6 +70,7 @@ cd ~/nccl/ make -j src.build NVCC_GENCODE="-gencode=arch=compute_121,code=sm_121" ## Set environment variables +export CUDA_HOME="/usr/local/cuda" export MPI_HOME="/usr/lib/aarch64-linux-gnu/openmpi" export NCCL_HOME="$HOME/nccl/build/" export LD_LIBRARY_PATH="$NCCL_HOME/lib:$CUDA_HOME/lib64/:$MPI_HOME/lib:$LD_LIBRARY_PATH" diff --git a/nvidia/trt-llm/README.md b/nvidia/trt-llm/README.md index d2de4ba..a17ed2c 100644 --- a/nvidia/trt-llm/README.md +++ b/nvidia/trt-llm/README.md @@ -97,7 +97,6 @@ The following models are supported with TensorRT-LLM on Spark. All listed models | **Qwen3-8B** | NVFP4 | ✅ | `nvidia/Qwen3-8B-FP4` | | **Qwen3-14B** | FP8 | ✅ | `nvidia/Qwen3-14B-FP8` | | **Qwen3-14B** | NVFP4 | ✅ | `nvidia/Qwen3-14B-FP4` | -| **Qwen3-32B** | NVFP4 | ✅ | `nvidia/Qwen3-32B-FP4` | | **Phi-4-multimodal-instruct** | FP8 | ✅ | `nvidia/Phi-4-multimodal-instruct-FP8` | | **Phi-4-multimodal-instruct** | NVFP4 | ✅ | `nvidia/Phi-4-multimodal-instruct-FP4` | | **Phi-4-reasoning-plus** | FP8 | ✅ | `nvidia/Phi-4-reasoning-plus-FP8` |