You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
(base) jovyan@ck-test-vscode-notebook-gpu-ok-0:~$ nvidia-smi
bash: nvidia-smi: command not found
(base) jovyan@ck-test-vscode-notebook-gpu-ok-0:~$ export PATH=$PATH:/usr/local/nvidia/bin
(base) jovyan@ck-test-vscode-notebook-gpu-ok-0:~$ nvidia-smi
NVIDIA-SMI couldn't find libnvidia-ml.so library in your system. Please make sure that the NVIDIA Display Driver is properly installed and present in your system.
Please also try adding directory that contains libnvidia-ml.so to your system PATH.
(base) jovyan@ck-test-vscode-notebook-gpu-ok-0:~$ export LD_LIBRARY_PATH=/usr/local/nvidia/lib:/usr/local/nvidia/lib64
(base) jovyan@ck-test-vscode-notebook-gpu-ok-0:~$ nvidia-smi
Tue Jan 7 02:48:50 2025
+-----------------------------------------------------------------------------------------+
| NVIDIA-SMI 560.35.03 Driver Version: 560.35.03 CUDA Version: 12.6 |
|-----------------------------------------+------------------------+----------------------+
| GPU Name Persistence-M | Bus-Id Disp.A | Volatile Uncorr. ECC |
| Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. |
| | | MIG M. |
|=========================================+========================+======================|
| 0 Tesla T4 Off | 00000000:00:05.0 Off | 0 |
| N/A 37C P8 10W / 70W | 1MiB / 15360MiB | 0% Default |
| | | N/A |
+-----------------------------------------+------------------------+----------------------+
+-----------------------------------------------------------------------------------------+
| Processes: |
| GPU GI CI PID Type Process name GPU Memory |
| ID ID Usage |
|=========================================================================================|
| No running processes found |
+-----------------------------------------------------------------------------------------+
Note: this issue is also present in jupyter-pytorch-cuda-full
The text was updated successfully, but these errors were encountered:
MeCode4Food
changed the title
GKE GPU nodes: nvidia-smi not found, likely missingPATH and LD_LIBRARY_PATH
GKE GPU nodes: nvidia-smi not found, likely missing env PATH and LD_LIBRARY_PATHJan 7, 2025
LD_LIBRARY_PATH can be easily added via the PodDefaults CRD, but PATH variable is not easily extended with that. Creating a new image to extend the PATH variable resolves this issue, but I wonder if this can be added into the base image(s), unless there is a way to resolve this without.
I am trying to run a vscode notebook on my GKE cluster's kubeflow platform. On the node(s), nvidia-drivers are already installed:
However, in the notebook this is not the case
This is however remedied with a change to
PATH
Note: this issue is also present in
jupyter-pytorch-cuda-full
The text was updated successfully, but these errors were encountered: