R
RunPod2mo ago
Emad

Can't run a 70B Llama 3.1 model on 2 A100 80 gb GPUs.

Hey, so I tired running the 70B llama model on 2gpu/worker but it keeps getting stuck at the same place every time but instead if I switch to the 8B model on 1 gpu/worker with a 48gb GPU, it works easily. The issue is coming with the 70B paramater model on 2 gpus/worker.
37 Replies
nerdylive
nerdylive2mo ago
Maybe 70b needs 192gbs or smth like that
Emad
Emad2mo ago
RunPod Blog
Run Larger LLMs on RunPod Serverless Than Ever Before - Llama-3 70B...
Up until now, RunPod has only supported using a single GPU in Serverless, with the exception of using two 48GB cards (which honestly didn't help, given the overhead involved in multi-GPU setups for LLMs.) You were effectively limited to what you could fit in 80GB, so you would essentially be
Emad
Emad2mo ago
This blogpost said that 2 80GB are enough
nerdylive
nerdylive2mo ago
yeah im not sure about the minimum requirements, maybe let me check
Emad
Emad2mo ago
alright also how much network volume do you I think need for this?
nerdylive
nerdylive2mo ago
maybe around 150~
Emad
Emad2mo ago
alright thanks let me know about the requirements
nerdylive
nerdylive2mo ago
can you try other, gpu 4x
Emad
Emad2mo ago
alr lemme try that 4090?
nerdylive
nerdylive2mo ago
4x 48 gb srry*
Emad
Emad2mo ago
ok np
Emad
Emad2mo ago
It got stuck here again
Emad
Emad2mo ago
It's always at this place What do you think could be the problem @nerdylive
Emad
Emad2mo ago
It went a bit further now
No description
Emad
Emad2mo ago
and now it just shifted to a different worker
Want results from more Discord servers?
Add your server