R
RunPod2mo ago
Yobin

Can you now run gemma 3 in the vllm container?

In the serverless, its seems im getting an error, any help on this
18 Replies
Jason
Jason2mo ago
can you send the error
Yobin
YobinOP2mo ago
I deleted it but it seems bc gemma3 is a new model so transformer is relatively outdated afaik?
Jason
Jason2mo ago
hmm did you use vllm? or only transformer? yeah maybe its a good thing to check first for compatibility
Yobin
YobinOP2mo ago
I used the preset vllm, llama 3.2b worked but the new gemma 3 didnt
Dj
Dj2mo ago
vLLM needs to publish an update first unfortunately You can use vLLM directly from the main branch, but that's not super easy if you're using our vLLM template iirc
Jason
Jason2mo ago
I think we can update and build our own vllm template from that vllm-worker repo in github easily Just update the requirement.txt or wherever installs the vllm
Bj9000
Bj90002mo ago
Looks like vllm v0.8.0 added gemma3 support will the serverless vllm be updated soon?
Jason
Jason2mo ago
usually its delayed, so probably a few days/ weeks late
Aizen
Aizen4w ago
Hi, i have the same issue , have you resolved it? , if then please help me out with it too
Yobin
YobinOP4w ago
I used ollana Ollama
Aizen
Aizen4w ago
Okay
Jason
Jason4w ago
yes i think vllm is updated already
Javier
Javier4w ago
I deployed an end point to try to call gemma3:4b but nothing is happening when I call it, anybody managed?
Jason
Jason4w ago
Yes it works
Jason
Jason4w ago
just now i've tried for you!
Jason
Jason4w ago
you need access from hf + hf token to access it
Jason
Jason4w ago
use vllm to configure + check the allow remote code options in the config ( in runpod menu when configuring vllm )
No description
Jason
Jason4w ago
second image is the next page
No description
No description

Did you find this page helpful?