Can multiple models be queried using the vllm serverless worker?

Just getting started with the vllm serverless worker and my first question is can I query multiple models via a single vllm serverless endpoint, or is it only possible to query one model per endpoint? If mutiple is possible are there any special steps to get it to work? If the answer to the above is no (it's 1 endpoint per model), is it recommended to use a 1 to 1 serverless endpoint to network volume? FYI, I tried pre-loading some models onto my volume, but my serverless endpoint could not find any of them other than the one explicitly loaded via the vllm MODEL_NAME env var, so not sure if I'm just missing something or that is a limitation.
1 Reply
justin
justin10mo ago
@Alpay Ariyak Might be the best to answer this
Want results from more Discord servers?
Add your server