Pod with multiple gpus (rtx 4090)
When I am starting runpod/pytorch:2.4.0-py3.11-cuda12.4.1-devel-ubuntu22.04 with multiple gpus (rtx 4090) in my simple .cu file I am just trying to get cudaGetDeviceCount and getting 999 unknown error. when running code on instance with one gpu all is working fine. in nvidia-smi all gpus are visible. maybe i am missing anything obvious in pod setup?
1 Reply
well.. now all works fine.. maybe previously i got some instance with old drivers or don't know why now everything is working..