Collection of step-by-step playbooks for setting up AI/ML workloads on NVIDIA DGX Spark devices with Blackwell architecture.
Go to file
Santosh Bhavani 0d5b85cdc5 Optimize Docker build with multi-stage caching
- Implement multi-stage Dockerfile (deps → builder → runner)
- Add BuildKit cache mounts for pnpm store and Next.js build cache
- Enable Next.js standalone output for smaller production images
- Create non-root user (nextjs:nodejs) with proper permissions
- Enhance .dockerignore to exclude more build artifacts
- Build time reduced from 225+ seconds to ~35 seconds

🤖 Generated with [Claude Code](https://claude.com/claude-code)

Co-Authored-By: Claude <noreply@anthropic.com>
2025-10-25 13:48:36 -07:00
nvidia Optimize Docker build with multi-stage caching 2025-10-25 13:48:36 -07:00
src/images chore: Regenerate all playbooks 2025-10-03 20:46:11 +00:00
LICENSE chore: Regenerate all playbooks 2025-10-03 20:46:11 +00:00
README.md chore: Regenerate all playbooks 2025-10-18 21:28:42 +00:00

NVIDIA DGX Spark

DGX Spark Playbooks

Collection of step-by-step playbooks for setting up AI/ML workloads on NVIDIA DGX Spark devices with Blackwell architecture.

About

These playbooks provide detailed instructions for:

  • Installing and configuring popular AI frameworks
  • Running inference with optimized models
  • Setting up development environments
  • Connecting and managing your DGX Spark device

Each playbook includes prerequisites, step-by-step instructions, troubleshooting guidance, and example code.

Available Playbooks

NVIDIA

Resources

License

See: