From 9414a5141f77965e6946f8273def846302e96d3d Mon Sep 17 00:00:00 2001 From: GitLab CI Date: Tue, 7 Apr 2026 04:13:30 +0000 Subject: [PATCH] chore: Regenerate all playbooks --- nvidia/txt2kg/assets/deploy/compose/docker-compose.yml | 4 ++-- nvidia/txt2kg/assets/frontend/lib/text-processor.ts | 2 +- 2 files changed, 3 insertions(+), 3 deletions(-) diff --git a/nvidia/txt2kg/assets/deploy/compose/docker-compose.yml b/nvidia/txt2kg/assets/deploy/compose/docker-compose.yml index b099b40..37981b7 100644 --- a/nvidia/txt2kg/assets/deploy/compose/docker-compose.yml +++ b/nvidia/txt2kg/assets/deploy/compose/docker-compose.yml @@ -27,8 +27,8 @@ services: # Ollama configuration - OLLAMA_BASE_URL=http://ollama:11434/v1 - OLLAMA_MODEL=llama3.1:8b - # Disable vLLM - - VLLM_BASE_URL=http://localhost:8001/v1 + # vLLM disabled in default Ollama mode + # - VLLM_BASE_URL=http://localhost:8001/v1 - VLLM_MODEL=disabled # Vector DB configuration - QDRANT_URL=http://qdrant:6333 diff --git a/nvidia/txt2kg/assets/frontend/lib/text-processor.ts b/nvidia/txt2kg/assets/frontend/lib/text-processor.ts index 21c8cab..bdac21c 100644 --- a/nvidia/txt2kg/assets/frontend/lib/text-processor.ts +++ b/nvidia/txt2kg/assets/frontend/lib/text-processor.ts @@ -108,7 +108,7 @@ export class TextProcessor { // Determine which LLM provider to use based on configuration // Priority: vLLM > NVIDIA > Ollama - if (process.env.VLLM_BASE_URL) { + if (process.env.VLLM_BASE_URL && process.env.VLLM_MODEL && process.env.VLLM_MODEL !== 'disabled') { this.selectedLLMProvider = 'vllm'; } else if (process.env.NVIDIA_API_KEY) { this.selectedLLMProvider = 'nvidia';