mirror of
https://github.com/NVIDIA/dgx-spark-playbooks.git
synced 2026-04-23 18:33:54 +00:00
chore: Regenerate all playbooks
This commit is contained in:
parent
9414a5141f
commit
8452a1c5b1
@ -172,12 +172,15 @@ Verify the NVIDIA runtime works:
|
|||||||
docker run --rm --runtime=nvidia --gpus all ubuntu nvidia-smi
|
docker run --rm --runtime=nvidia --gpus all ubuntu nvidia-smi
|
||||||
```
|
```
|
||||||
|
|
||||||
If you get a permission denied error on `docker`, add your user to the Docker group and log out/in:
|
If you get a permission denied error on `docker`, add your user to the Docker group and activate the new group in your current session:
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
sudo usermod -aG docker $USER
|
sudo usermod -aG docker $USER
|
||||||
|
newgrp docker
|
||||||
```
|
```
|
||||||
|
|
||||||
|
This applies the group change immediately. Alternatively, you can log out and back in instead of running `newgrp docker`.
|
||||||
|
|
||||||
> [!NOTE]
|
> [!NOTE]
|
||||||
> DGX Spark uses cgroup v2. OpenShell's gateway embeds k3s inside Docker and needs host cgroup namespace access. Without `default-cgroupns-mode: host`, the gateway can fail with "Failed to start ContainerManager" errors.
|
> DGX Spark uses cgroup v2. OpenShell's gateway embeds k3s inside Docker and needs host cgroup namespace access. Without `default-cgroupns-mode: host`, the gateway can fail with "Failed to start ContainerManager" errors.
|
||||||
|
|
||||||
@ -322,13 +325,21 @@ http://127.0.0.1:18789/#token=<long-token-here>
|
|||||||
|
|
||||||
**If accessing the Web UI from a remote machine**, you need to set up port forwarding.
|
**If accessing the Web UI from a remote machine**, you need to set up port forwarding.
|
||||||
|
|
||||||
|
First, find your Spark's IP address. On the Spark, run:
|
||||||
|
|
||||||
|
```bash
|
||||||
|
hostname -I | awk '{print $1}'
|
||||||
|
```
|
||||||
|
|
||||||
|
This prints the primary IP address (e.g. `192.168.1.42`). You can also find it in **Settings > Wi-Fi** or **Settings > Network** on the Spark's desktop, or check your router's connected-devices list.
|
||||||
|
|
||||||
Start the port forward on the Spark host:
|
Start the port forward on the Spark host:
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
openshell forward start 18789 my-assistant --background
|
openshell forward start 18789 my-assistant --background
|
||||||
```
|
```
|
||||||
|
|
||||||
Then from your remote machine, create an SSH tunnel to the Spark:
|
Then from your remote machine, create an SSH tunnel to the Spark (replace `<your-spark-ip>` with the IP address from above):
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
ssh -L 18789:127.0.0.1:18789 <your-user>@<your-spark-ip>
|
ssh -L 18789:127.0.0.1:18789 <your-user>@<your-spark-ip>
|
||||||
|
|||||||
Loading…
Reference in New Issue
Block a user