From 7bc85ebcc96419aa8e14898853a1a4e1cca35adc Mon Sep 17 00:00:00 2001 From: GitLab CI Date: Thu, 9 Oct 2025 15:38:30 +0000 Subject: [PATCH] chore: Regenerate all playbooks --- README.md | 6 +++--- nvidia/connect-to-your-spark/README.md | 4 ++-- nvidia/dgx-dashboard/README.md | 2 +- nvidia/flux-finetuning/README.md | 2 +- nvidia/jax/README.md | 2 +- nvidia/nvfp4-quantization/README.md | 2 +- nvidia/open-webui/README.md | 2 +- nvidia/speculative-decoding/README.md | 2 +- nvidia/vscode/README.md | 2 +- 9 files changed, 12 insertions(+), 12 deletions(-) diff --git a/README.md b/README.md index b51886b..4de7ea8 100644 --- a/README.md +++ b/README.md @@ -22,7 +22,7 @@ Each playbook includes prerequisites, step-by-step instructions, troubleshooting ### NVIDIA - [Comfy UI](nvidia/comfy-ui/) -- [Connect to Your Spark from Another Computer](nvidia/connect-to-your-spark/) +- [Set Up Local Network Access](nvidia/connect-to-your-spark/) - [DGX Dashboard](nvidia/dgx-dashboard/) - [FLUX.1 Dreambooth LoRA Fine-tuning](nvidia/flux-finetuning/) - [Optimized JAX](nvidia/jax/) @@ -33,9 +33,9 @@ Each playbook includes prerequisites, step-by-step instructions, troubleshooting - [NCCL for Two Sparks](nvidia/nccl/) - [Fine-tune with NeMo](nvidia/nemo-fine-tune/) - [Use a NIM on Spark](nvidia/nim-llm/) -- [Quantize to NVFP4](nvidia/nvfp4-quantization/) +- [NVFP4 Quantization](nvidia/nvfp4-quantization/) - [Ollama](nvidia/ollama/) -- [Use Open WebUI with Ollama](nvidia/open-webui/) +- [Open WebUI with Ollama](nvidia/open-webui/) - [Use Open Fold](nvidia/protein-folding/) - [Fine tune with Pytorch](nvidia/pytorch-fine-tune/) - [RAG application in AI Workbench](nvidia/rag-ai-workbench/) diff --git a/nvidia/connect-to-your-spark/README.md b/nvidia/connect-to-your-spark/README.md index 908d949..2306896 100644 --- a/nvidia/connect-to-your-spark/README.md +++ b/nvidia/connect-to-your-spark/README.md @@ -1,6 +1,6 @@ -# Connect to Your Spark from Another Computer +# Set Up Local Network Access -> Use NVIDIA Sync or manual SSH to connect to your Spark +> NVIDIA Sync helps set up and configure SSH access ## Table of Contents diff --git a/nvidia/dgx-dashboard/README.md b/nvidia/dgx-dashboard/README.md index 9504437..3be8cd5 100644 --- a/nvidia/dgx-dashboard/README.md +++ b/nvidia/dgx-dashboard/README.md @@ -1,6 +1,6 @@ # DGX Dashboard -> Manage your DGX system and launch JupyterLab +> Monitor your DGX system and launch JupyterLab ## Table of Contents diff --git a/nvidia/flux-finetuning/README.md b/nvidia/flux-finetuning/README.md index 93ddde1..71d6e6a 100644 --- a/nvidia/flux-finetuning/README.md +++ b/nvidia/flux-finetuning/README.md @@ -1,6 +1,6 @@ # FLUX.1 Dreambooth LoRA Fine-tuning -> Fine-tune FLUX.1-dev 12B model using multi-concept Dreambooth LoRA for custom image generation +> Fine-tune FLUX.1-dev 12B model using Dreambooth LoRA for custom image generation ## Table of Contents diff --git a/nvidia/jax/README.md b/nvidia/jax/README.md index 450d4e1..f979c26 100644 --- a/nvidia/jax/README.md +++ b/nvidia/jax/README.md @@ -1,6 +1,6 @@ # Optimized JAX -> Develop with Optimized JAX +> Optimize JAX to Run on Spark ## Table of Contents diff --git a/nvidia/nvfp4-quantization/README.md b/nvidia/nvfp4-quantization/README.md index 4ca1750..42bc6b4 100644 --- a/nvidia/nvfp4-quantization/README.md +++ b/nvidia/nvfp4-quantization/README.md @@ -1,4 +1,4 @@ -# Quantize to NVFP4 +# NVFP4 Quantization > Quantize a model to NVFP4 to run on Spark using TensorRT Model Optimizer diff --git a/nvidia/open-webui/README.md b/nvidia/open-webui/README.md index f02ebb6..3b4872b 100644 --- a/nvidia/open-webui/README.md +++ b/nvidia/open-webui/README.md @@ -1,4 +1,4 @@ -# Use Open WebUI with Ollama +# Open WebUI with Ollama > Install Open WebUI and use Ollama to chat with models on your Spark diff --git a/nvidia/speculative-decoding/README.md b/nvidia/speculative-decoding/README.md index 9b7474e..93b238b 100644 --- a/nvidia/speculative-decoding/README.md +++ b/nvidia/speculative-decoding/README.md @@ -1,6 +1,6 @@ # Speculative Decoding -> Learn how to setup speculative decoding for fast inference on Spark +> Learn how to set up speculative decoding for fast inference on Spark ## Table of Contents diff --git a/nvidia/vscode/README.md b/nvidia/vscode/README.md index 26d2e0d..a407656 100644 --- a/nvidia/vscode/README.md +++ b/nvidia/vscode/README.md @@ -1,6 +1,6 @@ # Install VS Code -> Install and use VS Code locally or remotely on Spark +> Install and use VS Code locally or remotely ## Table of Contents