HTTP 502 on VLLM pod
I'm getting a 502 when trying to connect to the deployed service. Using the vllm-latest image and these arguments:
--host 0.0.0.0 --port 8000 --model mistralai/Mistral-Small-24B-Instruct-2501 --dtype auto --enforce-eager --gpu-memory-utilization 0.95 --tensor-parallel-size 2
Using the ollama service doesn't have any issues. Any ideas?
1 Reply
is your logs for vllm telling its ready?
port 8000 http correct?
502 usually means app isnt ready
A 502 error indicates that Cloudflare received an invalid response from the origin server, causing the request to fail. Understanding the root causes and solutions for this error helps maintain the website’s availability and reliability.