Mihály
Mihály
RRunPod
Created by Mihály on 10/17/2024 in #⚡|serverless
Worker keeps running after finishing job, burning money?
No description
5 replies
RRunPod
Created by vitalik on 10/10/2024 in #⚡|serverless
Job retry after successful run
For me, upgrading the SDK from 1.7.1 to ,1.7.2 got rid of the retries
8 replies
RRunPod
Created by rougsig on 9/27/2024 in #⚡|serverless
Stuck IN_PROGRESS but job completed and worker exited
Ah, makes sense!
17 replies
RRunPod
Created by rougsig on 9/27/2024 in #⚡|serverless
Stuck IN_PROGRESS but job completed and worker exited
I'm not sure i follow 😄
17 replies
RRunPod
Created by rougsig on 9/27/2024 in #⚡|serverless
Stuck IN_PROGRESS but job completed and worker exited
Yeah, its noxhy2en39n3y3 my dev endpoint.
17 replies
RRunPod
Created by rougsig on 9/27/2024 in #⚡|serverless
Stuck IN_PROGRESS but job completed and worker exited
Hey @deanQ I've upgraded to 1.7.2 (5-6 hours ago) but still getting stuck jobs the same way. 😦
17 replies
RRunPod
Created by Untrack4d on 9/30/2024 in #⚡|serverless
Callback Function
Yep, i think this is what you are looking for https://docs.runpod.io/serverless/endpoints/send-requests#webhooks
3 replies
RRunPod
Created by rougsig on 9/27/2024 in #⚡|serverless
Stuck IN_PROGRESS but job completed and worker exited
On my side, it was added to the code after the above issue started happening. Didn't affect the outcomes
17 replies
RRunPod
Created by rougsig on 9/27/2024 in #⚡|serverless
Stuck IN_PROGRESS but job completed and worker exited
I dont know if that helps, but the 1.7.1 is more verbose with the errors as well :
"{"trace_id": "4453ebf4-f262-4e44-a422-d6f3691ac250", "request_id": "7b2b6fc3-2968-407c-bb29-97a14b07f238-e1", "user_agent": "RunPod-Python-SDK/1.7.1 (Linux 6.2.0-34-generic; x86_64) Language/Python 3.10.12", "start_time": "2024-09-29T15:54:28.322649+00:00", "method": "GET", "url": "https://api.runpod.ai/v2/74jm2u3liu0pcy/job-take/li0g2epzy6h0eu?gpu=NVIDIA GeForce RTX 4090&job_in_progress=0", "mode": "async", "connect": 0.2, "payload_size_bytes": 0, "exception": "", "transfer": 812777.8, "end_time": "2024-09-29T16:08:01.100656+00:00", "total": 812778.0}"
"{"trace_id": "4453ebf4-f262-4e44-a422-d6f3691ac250", "request_id": "7b2b6fc3-2968-407c-bb29-97a14b07f238-e1", "user_agent": "RunPod-Python-SDK/1.7.1 (Linux 6.2.0-34-generic; x86_64) Language/Python 3.10.12", "start_time": "2024-09-29T15:54:28.322649+00:00", "method": "GET", "url": "https://api.runpod.ai/v2/74jm2u3liu0pcy/job-take/li0g2epzy6h0eu?gpu=NVIDIA GeForce RTX 4090&job_in_progress=0", "mode": "async", "connect": 0.2, "payload_size_bytes": 0, "exception": "", "transfer": 812777.8, "end_time": "2024-09-29T16:08:01.100656+00:00", "total": 812778.0}"
"payload_size_bytes": 0 <-- seems sus?
17 replies
RRunPod
Created by rougsig on 9/27/2024 in #⚡|serverless
Stuck IN_PROGRESS but job completed and worker exited
No description
17 replies
RRunPod
Created by rougsig on 9/27/2024 in #⚡|serverless
Stuck IN_PROGRESS but job completed and worker exited
I'm having the same issue. SDK 1.6.2, 1.7.0 and 1.7.1 all produces this, however very rarely. Sometimes 1 our of 6, sometimes 1 out of 30. Re-submitting the same payload can run without issues a second time.
17 replies
RRunPod
Created by Mihály on 9/10/2024 in #⚡|serverless
Jobs in queue for a long time, even when there is a worker available
No description
12 replies
RRunPod
Created by Mihály on 9/10/2024 in #⚡|serverless
Jobs in queue for a long time, even when there is a worker available
No description
12 replies
RRunPod
Created by Mihály on 9/10/2024 in #⚡|serverless
Jobs in queue for a long time, even when there is a worker available
Thank you, will do!
12 replies
RRunPod
Created by Mihály on 9/10/2024 in #⚡|serverless
Jobs in queue for a long time, even when there is a worker available
Sure! It was either noxh y2en 39n3 y3 or k5hi ftra iqq8 dw
12 replies
RRunPod
Created by Mihály on 9/10/2024 in #⚡|serverless
Jobs in queue for a long time, even when there is a worker available
No description
12 replies
RRunPod
Created by Mihály on 3/26/2024 in #⚡|serverless
Failed to return job results.
Update For later readers with similar problem: In this specific case, i have found the root of the issue. It was indeed the response size, however it was not exceeded when returning the final response, but when using the runpod.serverless.progress_update Also, in an email conversation, the following was pointer out: - "limit for /runsync is 20mb" - "/run has an even smaller limit for 10mb at the moment" - "its both the input and the output payload limit"
8 replies
RRunPod
Created by Mihály on 3/26/2024 in #⚡|serverless
Failed to return job results.
ps: Also, i'm uploading the results in code, and cleaning the response of them to avoid this, but still.
8 replies
RRunPod
Created by Mihály on 3/26/2024 in #⚡|serverless
Failed to return job results.
No description
8 replies