mirror of
https://github.com/NVIDIA/dgx-spark-playbooks.git
synced 2026-04-23 02:23:53 +00:00
- Integrate NVIDIA API as alternative to Ollama for graph queries - Implement thinking tokens API with /think system message - Add min_thinking_tokens (1024) and max_thinking_tokens (2048) - Format reasoning_content with <think> tags for UI parsing - Support dynamic model/provider selection per query - Maintain Ollama fallback for backward compatibility This enables Traditional Graph to use NVIDIA's reasoning models (e.g., nvidia-nemotron-nano-9b-v2) with visible chain-of-thought. 🤖 Generated with [Claude Code](https://claude.com/claude-code) Co-Authored-By: Claude <noreply@anthropic.com> |
||
|---|---|---|
| .. | ||
| comfy-ui | ||
| connect-to-your-spark | ||
| connect-two-sparks | ||
| cuda-x-data-science | ||
| dgx-dashboard | ||
| flux-finetuning | ||
| jax | ||
| llama-factory | ||
| multi-agent-chatbot | ||
| multi-modal-inference | ||
| nccl | ||
| nemo-fine-tune | ||
| nim-llm | ||
| nvfp4-quantization | ||
| ollama | ||
| open-webui | ||
| pytorch-fine-tune | ||
| rag-ai-workbench | ||
| speculative-decoding | ||
| tailscale | ||
| trt-llm | ||
| txt2kg | ||
| unsloth | ||
| vibe-coding | ||
| vllm | ||
| vlm-finetuning | ||
| vscode | ||
| vss | ||