Someone is using my CUDA Memory?

Hi people, I get error when I try to train my model: torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 2.06 GiB. GPU 0 has a total capacty of 19.67 GiB of which 30.31 MiB is free. Process 2169311 has 19.64 GiB memory in use. Of the allocated memory 19.43 GiB is allocated by PyTorch, and 22.72 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting max_split_size_mb to avoid fragmentation. See documentation for Memory Management and PYTORCH_CUDA_ALLOC_CONF I do not really see how, since I just booted the server, and the task is relatively simple, how is my memory fully used?
5 Replies
ege0189
ege0189OP4w ago
This happens with every pod I have btw, I have tried closing, re-opening new pods, getting bigger GPUS, my VRAM's are all gone
nerdylive
nerdylive4w ago
can you send some of your past pod id's
Poddy
Poddy4w ago
@ege0189
Escalated To Zendesk
The thread has been escalated to Zendesk!
jurassimo
jurassimo3w ago
Yes, I have the same problem. Every pod which i wanna to start has less than 30% of memory.
nerdylive
nerdylive3w ago
which gpu? and try to press contact button on website, then create a new support request with your pod id ( check from audit log maybeif you had deleted/stop it)
Want results from more Discord servers?
Add your server