0% GPU utilization and 100% CPU utilization on Faster Whisper quick deploy endpoint
I used the "Quick Deploy" option to deploy a Faster Whisper custom endpoint (https://github.com/runpod-workers/worker-faster_whisper). Then, I called the endpoint to transcribe a 1 hour long podcast by using the following parameters:
The job completed in 201 seconds. I'm not sure if this is actually using the GPU and the graphs are wrong, or it's actually only using the CPU and it would have completed much faster had it been using the GPU.
4 Replies
Try checking the code to make sure it does uses the gpu
I think it does but to make sure just check the code, or launch a cpu instance with 20+ vcores and try
I am getting back "device": "cuda" in my output: https://github.com/runpod-workers/worker-faster_whisper/blob/main/src/predict.py#L120
Does that mean that it's actually using the GPU?
Yeah should be
Ok I think you're right, I tried it on a 32 cpu instance and got a bunch of
nvidia-smi: not found
logs plus it took longer than 200 seconds
So I guess the graph is wrong then
Thank you for your help!