NVIDIA DGX Spark

# DGX Spark Playbooks Collection of step-by-step playbooks for setting up AI/ML workloads on NVIDIA DGX Spark devices with Blackwell architecture. ## About These playbooks provide detailed instructions for: - Installing and configuring popular AI frameworks - Running inference with optimized models - Setting up development environments - Connecting and managing your DGX Spark device Each playbook includes prerequisites, step-by-step instructions, troubleshooting guidance, and example code. ## Use as Claude Code Skills Every playbook in this repo is also exposed as a [Claude Code](https://docs.claude.com/claude-code) skill, so Claude can install and configure them for you interactively. An index skill (`dgx-spark`) routes broad questions to the right specific playbook, and each leaf skill carries cross-references to related playbooks (prerequisites, alternatives, what to try next). ### Option A: install via Claude Code plugin marketplace (recommended) In Claude Code, run: ``` /plugin marketplace add jkneen/dgx-spark-playbooks /plugin install dgx-spark-playbooks@dgx-spark-playbooks ``` ### Option B: install locally via script Clone and run the install script. This symlinks each skill into `~/.claude/skills/`, so `git pull` updates them in place. ```bash git clone https://github.com/jkneen/dgx-spark-playbooks cd dgx-spark-playbooks ./scripts/install.sh # individual skills → ~/.claude/skills/ ./scripts/install.sh --plugin # or: whole repo as a plugin → ~/.claude/plugins/ ./scripts/uninstall.sh # remove ``` Requires Node 18+ if you want to regenerate skills from overrides; otherwise the committed `skills/` directory is used as-is. ### Customizing a skill Hand-curate any skill by creating or editing `overrides/.md` (see [`overrides/ollama.md`](overrides/ollama.md) as a worked example), then run `node scripts/generate.mjs` to rebuild. Overrides contribute the frontmatter `description` (which controls when the skill triggers) and any extra sections like "Related skills" or gotchas. Generated content between `` and `` markers is rewritten from the upstream README on every regeneration; override content outside those markers is preserved. A GitHub Action auto-regenerates `skills/` whenever a playbook README or override changes. ## Available Playbooks ### NVIDIA - [Comfy UI](nvidia/comfy-ui/) - [Connect Three DGX Spark in a Ring Topology](nvidia/connect-three-sparks/) - [Set Up Local Network Access](nvidia/connect-to-your-spark/) - [Connect Two Sparks](nvidia/connect-two-sparks/) - [CUDA-X Data Science](nvidia/cuda-x-data-science/) - [DGX Dashboard](nvidia/dgx-dashboard/) - [FLUX.1 Dreambooth LoRA Fine-tuning](nvidia/flux-finetuning/) - [Install and Use Isaac Sim and Isaac Lab](nvidia/isaac/) - [Optimized JAX](nvidia/jax/) - [Live VLM WebUI](nvidia/live-vlm-webui/) - [Run models with llama.cpp on DGX Spark](nvidia/llama-cpp/) - [LLaMA Factory](nvidia/llama-factory/) - [LM Studio on DGX Spark](nvidia/lm-studio/) - [Build and Deploy a Multi-Agent Chatbot](nvidia/multi-agent-chatbot/) - [Multi-modal Inference](nvidia/multi-modal-inference/) - [Connect Multiple DGX Spark through a Switch](nvidia/multi-sparks-through-switch/) - [NCCL for Two Sparks](nvidia/nccl/) - [Fine-tune with NeMo](nvidia/nemo-fine-tune/) - [NemoClaw with Nemotron 3 Super and Telegram on DGX Spark](nvidia/nemoclaw/) - [Nemotron-3-Nano with llama.cpp](nvidia/nemotron/) - [NIM on Spark](nvidia/nim-llm/) - [NVFP4 Quantization](nvidia/nvfp4-quantization/) - [Ollama](nvidia/ollama/) - [Open WebUI with Ollama](nvidia/open-webui/) - [OpenClaw 🦞](nvidia/openclaw/) - [Secure Long Running AI Agents with OpenShell on DGX Spark](nvidia/openshell/) - [Portfolio Optimization](nvidia/portfolio-optimization/) - [Fine-tune with Pytorch](nvidia/pytorch-fine-tune/) - [RAG Application in AI Workbench](nvidia/rag-ai-workbench/) - [SGLang for Inference](nvidia/sglang/) - [Single-cell RNA Sequencing](nvidia/single-cell/) - [Spark & Reachy Photo Booth](nvidia/spark-reachy-photo-booth/) - [Speculative Decoding](nvidia/speculative-decoding/) - [Set up Tailscale on Your Spark](nvidia/tailscale/) - [TRT LLM for Inference](nvidia/trt-llm/) - [Text to Knowledge Graph on DGX Spark](nvidia/txt2kg/) - [Unsloth on DGX Spark](nvidia/unsloth/) - [Vibe Coding in VS Code](nvidia/vibe-coding/) - [vLLM for Inference](nvidia/vllm/) - [VS Code](nvidia/vscode/) - [Build a Video Search and Summarization (VSS) Agent](nvidia/vss/) ## Resources - **Documentation**: https://www.nvidia.com/en-us/products/workstations/dgx-spark/ - **Developer Forum**: https://forums.developer.nvidia.com/c/accelerated-computing/dgx-spark-gb10 - **Terms of Service**: https://assets.ngc.nvidia.com/products/api-catalog/legal/NVIDIA%20API%20Trial%20Terms%20of%20Service.pdf ## License See: - [LICENSE](LICENSE) for licensing information. - [LICENSE-3rd-party](LICENSE-3rd-party) for third-party licensing information.