Sulove
Sulove
RRunPod
Created by Sulove on 11/23/2024 in #⚡|serverless
RunPod GPU Availability: Volume and Serverless Endpoint Compatibility
Hey everyone! Quick question about RunPod's GPU availability across different deployment types. I'm a bit confused about something: I created a volume in a data center where only a few GPU types were available. But when I'm setting up a serverless endpoint, I see I can select configs with up to 8 GPUs - including some that weren't available when I created my volume. Also noticed that GPU availability keeps fluctuating - sometimes showing low availability and sometimes none at all. So I'm wondering: 1. What happens if I pick a GPU type for my serverless endpoint that wasn't originally available in my volume's data center? 2. If I stick to only the GPUs that were available when creating my network volume, how does that work when those GPUs suddenly show low/no availability? Just trying to understand how RunPod handles these scenarios. Would really appreciate any insights! 🤔 Thanks in advance!"
3 replies