The Bloke LLM Template ExLlamaV2Cache_Q4 Error

Has anyone found a way around this. I use to use the pip install --upgrade exllamav2 command in the terminal but now that doesn't work. It worked yesterday but I'm guessing some things have changed and now it doesn't. The issue from what it seems has been going on for 2 or so weeks judging by the issues tab in github. https://github.com/TheBlokeAI/dockerLLM/issues/17. Using pip install --upgrade --no-deps exllamav2 solves it for now but that is only temporary I'm wondering if anyone has a updated version of the blokes template that can permanently solve this.
GitHub
Issues · TheBlokeAI/dockerLLM
TheBloke's Dockerfiles. Contribute to TheBlokeAI/dockerLLM development by creating an account on GitHub.
2 Replies
ashleyk
ashleyk10mo ago
TheBloke has disappeared for almost 2 months, you can use my template in the meantime.
Thick Thighs
Thick ThighsOP10mo ago
Thank you so much!
Want results from more Discord servers?
Add your server