R
RunPod9mo ago
J*

GPU for 13B language model

Just wanted to get your recommendations on GPU choice for running a 13B language model with a quantization in AWQ or GPTQ? Workload would be around 200-300 requests / hour. I tried a 48 GB A6000 with pretty good results but I was wondering if you think 24 GB GPU could be up to the task?
8 Replies
nerdylive
nerdylive9mo ago
Havent tried that yet, feel free to deploy it too
digigoblin
digigoblin9mo ago
24GB should be fine Best to try it and see
J*
J*OP9mo ago
Well I tried and failed, out of memory CUDA exception. I guess I'll stick to 48 GB GPU for now.
digigoblin
digigoblin9mo ago
Which model was it? Usually 13B with AWQ or GPTQ quantization aren't very large.
justin
justin9mo ago
Could be need to set some env variable sometimes for these LLMs to prevent it from eating up too much memory i think too. I remember i had a similar experience, but sometimes there are configs to help that If u wanna eat less memory
digigoblin
digigoblin9mo ago
Yeah for GPTQ, I had to set GPU_MEMORY_UTILIZATION to 0.80 instead of the default of 0.95
J*
J*OP9mo ago
Thanks, will give it a try!

Did you find this page helpful?