chore: Regenerate all playbooks

This commit is contained in:
GitLab CI 2025-10-05 00:47:50 +00:00
parent ea7e0b61c1
commit e139105ed3
2 changed files with 30 additions and 9 deletions

View File

@ -76,7 +76,6 @@ In a terminal, clone the repository and navigate to the flux-finetuning director
```bash ```bash
git clone https://gitlab.com/nvidia/dgx-spark/temp-external-playbook-assets/dgx-spark-playbook-assets/-/blob/main dgx-spark-playbooks git clone https://gitlab.com/nvidia/dgx-spark/temp-external-playbook-assets/dgx-spark-playbook-assets/-/blob/main dgx-spark-playbooks
cd dgx-spark-playbooks/nvidia/flux-finetuning
``` ```
## Step 3. Build the Docker container ## Step 3. Build the Docker container
@ -86,17 +85,21 @@ This docker image will download the required models and set up the environment f
- `ae.safetensors` - `ae.safetensors`
- `clip_l.safetensors` - `clip_l.safetensors`
- `t5xxl_fp16.safetensors` - `t5xxl_fp16.safetensors`
```bash ```bash
docker build -t flux-training . docker build -f Dockerfile.train --build-arg HF_TOKEN=$HF_TOKEN -t flux-training .
``` ```
## Step 4. Run the Docker container ## Step 4. Run the Docker container
```bash ```bash
## Run with GPU support and mount current directory ## Run with GPU support and mount current directory
docker run --gpus all -it --rm \ docker run -it \
-v $(pwd):/workspace \ --gpus all \
-p 8188:8188 \ --ipc=host \
--ulimit memlock=-1 \
--ulimit stack=67108864 \
--net=host \
flux-training flux-training
``` ```
@ -106,7 +109,7 @@ Inside the container, navigate to the sd-scripts directory and run the training
```bash ```bash
cd /workspace/sd-scripts cd /workspace/sd-scripts
../train.sh sh train.sh
``` ```
The training will: The training will:
@ -120,18 +123,37 @@ The training will:
After training completes, generate sample images: After training completes, generate sample images:
```bash ```bash
../inference.sh sh inference.sh
``` ```
This will generate several images demonstrating the learned concepts, stored in the `outputs` directory. This will generate several images demonstrating the learned concepts, stored in the `outputs` directory.
## Step 7. Spin up ComfyUI for visual workflows ## Step 7. Spin up ComfyUI for visual workflows
Build the Docker image for ComfyUI:
```bash
## Build the Docker image (this will download FLUX models automatically)
docker build -f Dockerfile.inference --build-arg HF_TOKEN=$HF_TOKEN -t flux-comfyui .
```
Run the ComfyUI container:
```bash
docker run -it \
--gpus all \
--ipc=host \
--ulimit memlock=-1 \
--ulimit stack=67108864 \
--net=host \
flux-comfyui
```
Start ComfyUI for an intuitive interface: Start ComfyUI for an intuitive interface:
```bash ```bash
cd /workspace/ComfyUI cd /workspace/ComfyUI
python main.py --listen 0.0.0.0 --port 8188 python main.py
``` ```
Access ComfyUI at `http://localhost:8188` Access ComfyUI at `http://localhost:8188`

View File

@ -84,7 +84,6 @@ In a terminal, clone the repository and navigate to the VLM fine-tuning director
```bash ```bash
git clone https://gitlab.com/nvidia/dgx-spark/temp-external-playbook-assets/dgx-spark-playbook-assets/-/blob/main dgx-spark-playbooks git clone https://gitlab.com/nvidia/dgx-spark/temp-external-playbook-assets/dgx-spark-playbook-assets/-/blob/main dgx-spark-playbooks
cd dgx-spark-playbooks/nvidia/vlm-finetuning
``` ```
## Step 3. Build the Docker container ## Step 3. Build the Docker container