How to respond to the requests at https://api.runpod.ai/v2/<YOUR ENDPOINT ID>/openai/v1

the openai input is in the job input, I extracted it and processes the request . when send the the response with yield or return it recived could you take a look at this [https://github.com/mohamednaji7/runpod-workers-scripts/blob/main/empty_test/test%20copy%203.py] [https://github.com/mohamednaji7/runpod-workers-scripts/blob/main/empty_test/handler.py] when I run made an request I got
raise HTTPError(http_error_msg, response=self)
requests.exceptions.HTTPError: 500 Server Error: Internal Server Error for url: https://api.runpod.ai/v2/nqe3wqry3h7noa/openai/v1/chat/completions
raise HTTPError(http_error_msg, response=self)
requests.exceptions.HTTPError: 500 Server Error: Internal Server Error for url: https://api.runpod.ai/v2/nqe3wqry3h7noa/openai/v1/chat/completions
4 Replies
nerdylive
nerdylive2d ago
Did you use an api key? if yes, what kind of permission did you select? oh wait.. you have to use a real model i think just test it with a real hf model that is supported by vllm
Mohamed Nagy
Mohamed NagyOP2d ago
I am doing a dummy test because I am building my worker and want to use unsloth instead of vllm and transformers I think its needs to route the /v1/chat/completion , waht do you think? I made a bunch of failed trials
nerdylive
nerdylive17h ago
I think the /openai/v1 endpoint is always exposed like in the vllm docs, you can try copying the input handler of vllm worker from runpod to try from there first In real endpoints* I'm not sure if it does the same in local
SvenBrnn
SvenBrnn7h ago
GitHub
runpod-worker-ollama/wrapper/src at master · SvenBrnn/runpod-worker...
A serverless ollama worker for runpod.io. Contribute to SvenBrnn/runpod-worker-ollama development by creating an account on GitHub.
GitHub
runpod-worker-ollama/test_inputs/openai_completion.json at master ·...
A serverless ollama worker for runpod.io. Contribute to SvenBrnn/runpod-worker-ollama development by creating an account on GitHub.
GitHub
runpod-worker-ollama/test_inputs/openai_get_models.json at master ·...
A serverless ollama worker for runpod.io. Contribute to SvenBrnn/runpod-worker-ollama development by creating an account on GitHub.

Did you find this page helpful?