chore: Regenerate all playbooks

This commit is contained in:
GitLab CI 2025-10-09 21:28:46 +00:00
parent 63975362f1
commit bf842ce358
4 changed files with 3400 additions and 11 deletions

View File

@ -33,7 +33,7 @@ Each playbook includes prerequisites, step-by-step instructions, troubleshooting
- [Multi-modal Inference](nvidia/multi-modal-inference/)
- [NCCL for Two Sparks](nvidia/nccl/)
- [Fine-tune with NeMo](nvidia/nemo-fine-tune/)
- [Use a NIM on Spark](nvidia/nim-llm/)
- [NIM on Spark](nvidia/nim-llm/)
- [NVFP4 Quantization](nvidia/nvfp4-quantization/)
- [Ollama](nvidia/ollama/)
- [Open WebUI with Ollama](nvidia/open-webui/)

File diff suppressed because it is too large Load Diff

File diff suppressed because one or more lines are too long

View File

@ -1,6 +1,6 @@
# Use a NIM on Spark
# NIM on Spark
> Run an LLM NIM on Spark
> Deploy a NIM on Spark
## Table of Contents
@ -19,17 +19,11 @@
### Basic idea
NVIDIA Inference Microservices (NIMs) provide optimized containers for deploying large language
models with simplified APIs. This playbook demonstrates how to run LLM NIMs on DGX Spark devices,
enabling GPU-accelerated inference through Docker containers. You'll set up authentication with
NVIDIA's registry, launch a containerized LLM service, and perform basic inference testing to
verify functionality.
NVIDIA NIM is containerized software for fast, reliable AI model serving and inference on NVIDIA GPUs. This playbook demonstrates how to run NIM microservices for LLMs on DGX Spark devices, enabling local GPU inference through a simple Docker workflow. You'll authenticate with NVIDIA's registry, launch the NIM inference microservice, and perform basic inference testing to verify functionality.
### What you'll accomplish
You'll deploy an LLM NIM container on your DGX Spark device, configure it for GPU acceleration,
and establish a working inference endpoint that responds to HTTP API calls with generated text
completions.
You'll launch a NIM container on your DGX Spark device to expose a GPU-accelerated HTTP endpoint for text completions.
### What to know before starting