18 Replies
I don't get this, what's the problem?
And also for logs please copy paste them or send it in a text file
Please
It charged for 5.3 hours
When the request wasnt that long
The billing shows that it was running for 5.3 hrs
What makes that?
I'd suggest using webchat or email support to resolve this out
The program just ran into an error
which can be seen from the logs
How can it run for 5.3 hours
i dont know, maybe something is running check your docker template or image
what is running in main thread
and there is some kind of timeout for workers, what do you set it to?
usually an hour or less default
Should I resolve this with the webchat?
yes
or by email
they better understand about the usages and billing of your account too
From what I see your worker was throwing error and was looping
and if not proper handled it will loop worker causing it to never stop.
You probably want to check code of your worker
And the server was on for 5 hours?
yes if you did not set it to end. I'm not sure how your worker is build
I have noticed this too. If your LLM throws CUDA run out of memory, it will loop forever, unless you cancel the job manually. It's very dangerous on an expensive GPU. I wished there was a better way to handle that.
Eh isn't there already a running timeout?
You can set them in your endpoint settings
Can you show us how? That would great. Thanks
Go to your endpoint settings and take a screenshot
Expand anything you can find
And take a screen shot, let me see the options please
Something called "execution timeout" I'm very sure
Ahh yes. I found it. Mine isn't enabled.
60 seconds is reasonable, right?
Depends on your specific endpoint.
yep depends on what are you doing. try to estimate whats the normal time from your graphs maybe?