andyh3118
andyh3118
RRunPod
Created by andyh3118 on 3/5/2024 in #⚡|serverless
Hi, is there currently an outage to Serverless API?
thanks. let me look into that. Could be the issue with ExllamaV2.
39 replies
RRunPod
Created by andyh3118 on 3/5/2024 in #⚡|serverless
Hi, is there currently an outage to Serverless API?
got it.
39 replies
RRunPod
Created by andyh3118 on 3/5/2024 in #⚡|serverless
Hi, is there currently an outage to Serverless API?
So it get to the handler, but stuck 😆
39 replies
RRunPod
Created by andyh3118 on 3/5/2024 in #⚡|serverless
Hi, is there currently an outage to Serverless API?
hmm. you are correct.
39 replies
RRunPod
Created by andyh3118 on 3/5/2024 in #⚡|serverless
Hi, is there currently an outage to Serverless API?
This is our handler code:
import logging

import runpod

from app.exllamav2_common import boot_engine, generate

logger = logging.getLogger(__name__)

async def handler(job: dict):
request_dict: dict = job.pop("input", {})

configs_dict = request_dict.copy()

full_response = ""
for full_response in generate(configs_dict):
yield {"text": full_response, "finished": False}

yield {"text": full_response, "finished": True}


boot_engine()

def concurrency_modifier(current_concurrency):
max_concurrency = 1
return max(0, max_concurrency - current_concurrency)

runpod.serverless.start({
"handler": handler,
"return_aggregate_stream": True,
})import logging

import runpod

from app.exllamav2_common import boot_engine, generate

logger = logging.getLogger(__name__)

async def handler(job: dict):
request_dict: dict = job.pop("input", {})

configs_dict = request_dict.copy()

full_response = ""
for full_response in generate(configs_dict):
yield {"text": full_response, "finished": False}

yield {"text": full_response, "finished": True}


boot_engine()

def concurrency_modifier(current_concurrency):
max_concurrency = 1
return max(0, max_concurrency - current_concurrency)

runpod.serverless.start({
"handler": handler,
"return_aggregate_stream": True,
})
import logging

import runpod

from app.exllamav2_common import boot_engine, generate

logger = logging.getLogger(__name__)

async def handler(job: dict):
request_dict: dict = job.pop("input", {})

configs_dict = request_dict.copy()

full_response = ""
for full_response in generate(configs_dict):
yield {"text": full_response, "finished": False}

yield {"text": full_response, "finished": True}


boot_engine()

def concurrency_modifier(current_concurrency):
max_concurrency = 1
return max(0, max_concurrency - current_concurrency)

runpod.serverless.start({
"handler": handler,
"return_aggregate_stream": True,
})import logging

import runpod

from app.exllamav2_common import boot_engine, generate

logger = logging.getLogger(__name__)

async def handler(job: dict):
request_dict: dict = job.pop("input", {})

configs_dict = request_dict.copy()

full_response = ""
for full_response in generate(configs_dict):
yield {"text": full_response, "finished": False}

yield {"text": full_response, "finished": True}


boot_engine()

def concurrency_modifier(current_concurrency):
max_concurrency = 1
return max(0, max_concurrency - current_concurrency)

runpod.serverless.start({
"handler": handler,
"return_aggregate_stream": True,
})
39 replies
RRunPod
Created by andyh3118 on 3/5/2024 in #⚡|serverless
Hi, is there currently an outage to Serverless API?
based on the logs, the requests are not getting to the handler.
39 replies
RRunPod
Created by andyh3118 on 3/5/2024 in #⚡|serverless
Hi, is there currently an outage to Serverless API?
same image / model
39 replies
RRunPod
Created by andyh3118 on 3/5/2024 in #⚡|serverless
Hi, is there currently an outage to Serverless API?
Ah.. it has been working for 2-3 weeks (we used it very actively)
39 replies
RRunPod
Created by andyh3118 on 3/5/2024 in #⚡|serverless
Hi, is there currently an outage to Serverless API?
(to indicate that it is broken?)
39 replies
RRunPod
Created by andyh3118 on 3/5/2024 in #⚡|serverless
Hi, is there currently an outage to Serverless API?
ok. any logs on your end that you can share?
39 replies
RRunPod
Created by andyh3118 on 3/5/2024 in #⚡|serverless
Hi, is there currently an outage to Serverless API?
Ah. sorry, I was wrong. It is not vLLM. We use our own exllama image.
39 replies
RRunPod
Created by andyh3118 on 3/5/2024 in #⚡|serverless
Hi, is there currently an outage to Serverless API?
Thanks!
39 replies
RRunPod
Created by andyh3118 on 3/5/2024 in #⚡|serverless
Hi, is there currently an outage to Serverless API?
No description
39 replies
RRunPod
Created by andyh3118 on 3/5/2024 in #⚡|serverless
Hi, is there currently an outage to Serverless API?
Workers are running
39 replies
RRunPod
Created by andyh3118 on 3/5/2024 in #⚡|serverless
Hi, is there currently an outage to Serverless API?
No description
39 replies
RRunPod
Created by andyh3118 on 3/5/2024 in #⚡|serverless
Hi, is there currently an outage to Serverless API?
No description
39 replies
RRunPod
Created by andyh3118 on 3/5/2024 in #⚡|serverless
Hi, is there currently an outage to Serverless API?
Things started to become unstable since weekend, and today is full outage for us...
39 replies
RRunPod
Created by andyh3118 on 3/5/2024 in #⚡|serverless
Hi, is there currently an outage to Serverless API?
I don't think input body is wrong though because the same service has been running smoothly for 2-3 weeks already.
39 replies
RRunPod
Created by andyh3118 on 3/5/2024 in #⚡|serverless
Hi, is there currently an outage to Serverless API?
We are using vLLM
39 replies
RRunPod
Created by andyh3118 on 3/5/2024 in #⚡|serverless
Hi, is there currently an outage to Serverless API?
This is my endpoint 1ifuoxegzxuhb4
39 replies