Best Mixtral/LLaMA2 LLM for code-writing, inference, 24 to 48 GB?

Good evening all you experts! I'm past the pain and suffering stage and into the finesse and finishing stage - what is the best class of models for doing basic inference and in particular formulating simple commands based on a set of simple rules, and which will fit into a 24 GB (or 48 GB if much better) runpod?
4 Replies
Unknown User
Unknown User13mo ago
Message Not Public
Sign In & Join Server To View
Stone Johnson
Stone JohnsonOP13mo ago
Thanks a ton! Will try it out
Alpay Ariyak
Alpay Ariyak13mo ago
OpenChat-3.5-1210 :)
Stone Johnson
Stone JohnsonOP13mo ago
Gonna try it! Do I understand that it is 7B size, so probably runs in 16 GB?
Want results from more Discord servers?
Add your server