From 36ac5b74eb2273d8e7f972276ff745efb73d4588 Mon Sep 17 00:00:00 2001 From: GitLab CI Date: Thu, 2 Apr 2026 22:45:52 +0000 Subject: [PATCH] chore: Regenerate all playbooks --- nvidia/vllm/README.md | 3 +-- 1 file changed, 1 insertion(+), 2 deletions(-) diff --git a/nvidia/vllm/README.md b/nvidia/vllm/README.md index 8a5bfb1..cd0e8d0 100644 --- a/nvidia/vllm/README.md +++ b/nvidia/vllm/README.md @@ -148,8 +148,7 @@ vllm serve ${HF_MODEL_HANDLE} To run models from Gemma 4 model family, (e.g. `google/gemma-4-31B-it`): ```bash docker run -it --gpus all -p 8000:8000 \ -vllm/vllm-openai:gemma4-cu130 \ -vllm serve ${HF_MODEL_HANDLE} +vllm/vllm-openai:gemma4-cu130 ${HF_MODEL_HANDLE} ``` Expected output should include: