R
RunPod12mo ago
Shware

The actual storage space of the network volume is wrong.

Hi, I encountered a problem about the network storage. I applied for 4TB of network storage and have used about 2.6TB so far. However, I noticed that on the pod management page, the displayed information for my pod indicates that the volume usage has reached 94%. When I attempted to write approximately 200 to 300GB more to reach 100%, I received a notification that I had reached my quota limit, which doesn't align with the 4TB of space I applied for. Could you help me identify the issue? Thanks!
Solution:
Hi, your team has helped me with this problem. The reason is that I have a lot of small files
Jump to solution
22 Replies
nerdylive
nerdylive12mo ago
are you sure its not compressed files then extracted? try to check them manually by command lines first while waiting for supports
Solution
Shware
Shware12mo ago
Hi, your team has helped me with this problem. The reason is that I have a lot of small files
nerdylive
nerdylive12mo ago
ah ye its not my team actually, im not from runpod
dagger
dagger11mo ago
@River Snow Similar issue with me. So, a lot of small files will caused a bugs on volume indicator? I have 1274318 files in my dataset folder. but total size only 3.3 GB, smaller than my volume size
justin
justin11mo ago
Just as a side note, sometimes what also happens is: if deleted using jupyter notebook it can be hiding in the trash-0. just sometimes i see sometimes it hiding in the /.Trash-0 directory b/c of deleting with jupyter labs notebook.
flash-singh
flash-singh11mo ago
network volume storage uses at least 64kb space per file, if your file is smaller it still uses that much and if you have lots of small files, it will add up, this is not a bug but intended due to distributed storage chunks
ashleyk
ashleyk11mo ago
/.Trash-0 is for container disk, its /workspace/.Trash-0 for volume disk.
justin
justin11mo ago
😮 WOW that is wild, i learnt a lot today - this is great to know
dagger
dagger11mo ago
@flash-singh Thanks, But I have a lot of small file datasets to train on. Is there a way to handle that rather than increase the network storage?
justin
justin11mo ago
maybe can move it to container storage? and then zip it to move back to volume network storage?
dagger
dagger11mo ago
Thanks for your hint, But it still has to be extracted for training purposes. And as my data grows, I have to increase the network storage limit several times which impacts costs.
flash-singh
flash-singh11mo ago
no way around it unless you zip it to store in network storage like @justin suggested
dagger
dagger11mo ago
Noted, thanks for your help
justin
justin11mo ago
But you have two storage locations? One in Container , one in network storage. So you can extract to Container, outside your /workspace, and then it won't be 64kb per file? And then anytime u need to update keep the one in network for zip And u can just point ur code to the files outside of /workspace Idk could be wrong just my thoughts - sounds like a network storage issue
dagger
dagger11mo ago
Thanks, but the problem is that extraction takes more than hours on runpod. The files number in the millions. Have no idea why runpod is slower for extracting than in my local device (m1, 8B ram)
justin
justin11mo ago
Oh interesting Hm. Maybe you can tarball it instead of zipping? That way u arent running it through a compression / decompression algorithm Edit: prob wont work after reading Ah just read ur post.. yes.. this is a lot of data
dagger
dagger11mo ago
I've tried and it's the same. I think cpu processing in runpod is slower because the process is shared with other users even in secure cloud (CMIIW) 🙏 So, I only use the GPU on the runpod, because it is dedicated for my runpod. Other than that, I process it locally for any cpu processing and upload it to RunPod because of the time difference.
justin
justin11mo ago
Makes sense
flash-singh
flash-singh11mo ago
0 gpu gives very little cpu, if thats what your using, cpu only pods are around the corner
dagger
dagger11mo ago
But I was using 1x3090 GPU
flash-singh
flash-singh11mo ago
that should give you many more cores with zip or tar, make sure to use multi core instead of single core its much faster that way
dagger
dagger11mo ago
Agreed, it should be like that, just a simple unzip. I have no idea why it was slower
Want results from more Discord servers?
Add your server