R
RunPod•3mo ago
houmie

Can we run aphrodite-engine on Serverless?

aphrodite-engine is a fork from vLLM and also supports exl2 format, which gives it a huge advantage. Are there any plans to support aphrodite-engine in future on RunPod's serverless offering? I believe currently aphrodite-engine is only supported as a single server on RunPod. Thanks
8 Replies
Madiator2011
Madiator2011•3mo ago
I mean nothing stops your from building own worker 🙂
houmie
houmie•3mo ago
How? lol
Madiator2011
Madiator2011•3mo ago
all workers are open source so you can have look at code and build own. Some coding and then packaging docker image
houmie
houmie•3mo ago
Seriously, I would be interested building my own queue and host aphrodite-engine on a single pod instead.
houmie
houmie•3mo ago
GitHub
GitHub - runpod-workers/worker-vllm: The RunPod worker template for...
The RunPod worker template for serving our large language model endpoints. Powered by vLLM. - runpod-workers/worker-vllm
houmie
houmie•3mo ago
Or any other you could recommend?
Madiator2011
Madiator2011•3mo ago
I do not know what aphrodite-engine is so cant tell but yes if it's fork ov vllm it will be good start point
nerdylive
nerdylive•3mo ago
Yes find your own source study them and that will help you understand, the triggers of runpod serverless is quite easy They all are documented on runpod docs
Want results from more Discord servers?
Add your server
More Posts
Template pytorch-1.13.1 lists cuda 11.7.1 version but is actually cuda 11.8?I tried running a model that requires pytorch-1.13.1 and 11.7 but it said the cuda version doesn't mIdle timeout not workingHi team. I'm setting my serverless endpoint with a Idle timeout for 180 seconds, but it's idleing oCan't connect to sfptHi, I can't access sftp. On my previous pod I could do it and I just swaped the ip and the port, butUnable to ssh onto my pod with the public key already on the runpod serverI am unable to ssh into the pod when using the command from runpod's site: ``` name .ssh % ssh rootIs serverless cost per worker or per GPU?I'm looking at serverless GPU options and when looking at 48 GB GPU it costs $0.00048/s. But is thaopenai compatible endpoint for custom serverless docker imagehow can I get openai compatible endpoint for my custom docker image in runpod serverless. I am tryiSecurely using serverless endpoints on the client-side?I have a use case where I'm doing a client-server webapp that uses serverless endpoints. In order toI wanna use comfyUI for a img2vid workflow - can I do this via the serverless serviceI already tried setting up a pod yesterday where I uploaded all the needed models and stuff but todaUsing network volume with serverlessI am running a stateless model within serverless to modify provided image. I am wondering if the netPython modules missing when pod is startingWhen starting a comfy-ui pod after some downtime, I get a lot of messages of the kind ``` Import t