worker-vllm build fails
I am getting the following error when building the new worker-vllm image with my model.
18 Replies
@Alpay Ariyak I'm not sure what the exact issue is here. Do you need a GPU to build the image?
I checked out commit
2b5b8dfb61e32d221bc8ce49f98ec74698154a6e
to get it working for now. Seems latest release is broken somehowIts a regression, was fixed previously.
Actually, I had to go even further back to get it working 😅
Just pushing my new serverless image now. Would love for this to be fixed so that I can upgrade
bfeb60c54eaad2eeffa9741ce7600eb30e573698
Checking this now
This has to do with vLLM's updates
Seems to only be affecting AWQ and potentially other quants
So this is why https://github.com/vllm-project/vllm/blob/929b4f2973ec6a53ea4f0f03d21147ef8b8278be/vllm/model_executor/weight_utils.py#L85-L122
Love the
# TODO(woosuk): Move this to other place.
lol
The function is not used in that file in any wayGitHub
vllm/vllm/model_executor/weight_utils.py at 929b4f2973ec6a53ea4f0f0...
A high-throughput and memory-efficient inference and serving engine for LLMs - vllm-project/vllm
Hey @Casper. could you try with image 0.3.1 as base
In this line instead of 0.3.0 https://github.com/runpod-workers/worker-vllm/blob/717343b0ad4d8a4ea76626c52b473619c646e30b/Dockerfile#L2
Pushed a new image, should work
My model is not quantized
I’ll try it later
Got you
Should fix either way
Getting same error
ImportError: libcuda.so.1: cannot open shared object file: No such file or directory
What’s the full error
Same as I pasted above
Are you sure you’re using 0.3.1?
Yeah it loads the new 0.3.1
[vllm-base 1/7] FROM docker.io/runpod/worker-vllm:base-0.3.1-cuda11.8.0
I'm on the latest commit
d91ccb866fc784b81a558f0da44041a020ba54e0
I see what's going on
I am building on a Macbook M2 btw
Just for reference
Pushing new base image
Done, try now
Rebuilding now, let's see
Build worked @Alpay Ariyak, thanks for fixing it
No problem, thanks for pointing it out!