Collection of step-by-step playbooks for setting up AI/ML workloads on NVIDIA DGX Spark devices with Blackwell architecture.
Go to file
Santosh Bhavani 8c1d2ae9f3 feat(docker): add vector search services and GPU configuration
- Add optional Pinecone and sentence-transformers services for vector search
- Configure NVIDIA GPU support with proper environment variables
- Add new environment variables for embeddings and Pinecone
- Add docker compose profiles to optionally enable vector-search
- Improve CUDA configuration for Ollama service
- Add pinecone-net network for service communication
2025-10-19 19:56:55 -07:00
nvidia feat(docker): add vector search services and GPU configuration 2025-10-19 19:56:55 -07:00
src/images chore: Regenerate all playbooks 2025-10-03 20:46:11 +00:00
LICENSE chore: Regenerate all playbooks 2025-10-03 20:46:11 +00:00
README.md chore: Regenerate all playbooks 2025-10-17 17:29:40 +00:00

NVIDIA DGX Spark

DGX Spark Playbooks

Collection of step-by-step playbooks for setting up AI/ML workloads on NVIDIA DGX Spark devices with Blackwell architecture.

About

These playbooks provide detailed instructions for:

  • Installing and configuring popular AI frameworks
  • Running inference with optimized models
  • Setting up development environments
  • Connecting and managing your DGX Spark device

Each playbook includes prerequisites, step-by-step instructions, troubleshooting guidance, and example code.

Available Playbooks

NVIDIA

Resources

License

See: