Thick Thighs
Thick Thighs
RRunPod
Created by Thick Thighs on 3/21/2024 in #⛅|pods
The Bloke LLM Template ExLlamaV2Cache_Q4 Error
Has anyone found a way around this. I use to use the pip install --upgrade exllamav2 command in the terminal but now that doesn't work. It worked yesterday but I'm guessing some things have changed and now it doesn't. The issue from what it seems has been going on for 2 or so weeks judging by the issues tab in github. https://github.com/TheBlokeAI/dockerLLM/issues/17. Using pip install --upgrade --no-deps exllamav2 solves it for now but that is only temporary I'm wondering if anyone has a updated version of the blokes template that can permanently solve this.
3 replies