Throttled
Hey π
Is there something I can do to prevent getting throttled? I see the availability for the GPU I selected is high, and I'm also not using any network disk, so I'm a bit confused what's exactly happening.
ID: ofjdhe4djh1k5t
13 Replies
ended up having to manually terminate the throttled worker, and after that it automatically spun up another worker that got my requests unstuck. Not sure why I had to do this manually to get my queue unstuck
Increase the max number of workers you have set.
in my use case I only ever want to run one worker at the same time, as I will be running a task about every 30 seconds that's not time sensitive. I wouldn't mind if a throttled worker that's stuck for several minutes gets automatically terminated and spins up another
I think increase the maximum amount of workers, but set the queue time under advance for like 1 minute or something
you can set the time it takes before a worker spins up a different worker
that way u only pay for active workers
There just some weird behaviors with a small maximum worker
ah yeah that sounds good. In my case I'm not really doing anything time-sensitive, so I'll probably just set it at like 5 minutes so my queue is not stuck forever.
Still would be nice if workers that are throttled for a long time are automatically terminated, but yeah this should work for now. Thanks!
bit confused what's happening here now
my queue is empty for a while now, and I see this
my idle timeout is low and active workers 0, so I don't quite understand what it's trying to do, with my queue already being empty for more than 5 minutes or so
xD yea dont worry about it
Runpod adds an additional 3 workers in the background to help with scaling issues but u arent paying for it
nor count against ur limit
i also see the same thing
i wish runpod explained whatever is going on here more but never had an issue with it
ah alright
I didn't have this before when sticking with 1 max worker, so got a bit confused when seeing this after increasing the max to 2
Yeah - is a weird thing they do when ur greater than 1 worker max. π€·, wish was something documented / explained but so be it
https://discord.com/channels/912829806415085598/1185822418053386340
I wish for it too
Utilizing network storage if your docker image is 20+ GBoh uh my docker image is definitely not 19GB π it's incredibly bloated because I had to go through a lot of hoops to get GPU support working with my tensorflow for some reason. Currently still building my thing as a POC, so haven't really slimmed down the image yet
max 1 worker we treat as development, and give you only 1 worker, most people want to ssh and debug, 2 or more are considered production workload and we add additional worker caches to help reduce throttled workers
haha i think under 20 gb is actually ok
i think once it exceeds like 30 gb / climbs around that regions gets bad