GPU FAQs¶
GPU processes are not visible while running nvidia-smi inside a pod¶
Q: When running the nvidia-smi command inside a GPU-utilizing pod, no GPU process information is visible in the full-card mode and vGPU mode.
A: Due to PID namespace isolation, GPU processes are not visible inside the Pod. To view GPU processes, you can use one of the following methods:
- Configure the workload using the GPU with
hostPID: trueto enable viewing PIDs on the host. - Run the
nvidia-smicommand in the driver pod of the gpu-operator to view processes. - Run the
chroot /run/nvidia/driver nvidia-smicommand on the host to view processes.