Henky!!
RRunPod
•Created by Netrve on 7/24/2024 in #⚡|serverless
How does the vLLM template provide an OAI route?
No matter if its a exl2 webserver, kcpp webserver or ollama webserver it would work best that way
8 replies
RRunPod
•Created by Netrve on 7/24/2024 in #⚡|serverless
How does the vLLM template provide an OAI route?
All we need is a 1:1 mapping of the webserver
8 replies
RRunPod
•Created by Netrve on 7/24/2024 in #⚡|serverless
How does the vLLM template provide an OAI route?
We all need the same thing
8 replies
RRunPod
•Created by Netrve on 7/24/2024 in #⚡|serverless
How does the vLLM template provide an OAI route?
Alright because the actual worker part i need would be insanely simplistic
8 replies
RRunPod
•Created by Netrve on 7/24/2024 in #⚡|serverless
How does the vLLM template provide an OAI route?
+1 making a template for KCPP is pointless without this
8 replies
load a new network volumen into a pod?
The easiest way to do it with Koboldcpp is https://koboldai.org/runpodcpp it will allow you to specify the models in the runpod variables directly without the need for persistent or network storage.
3 replies
my pod start very slow
@justin We do still actively maintain it so if people have issues feel free to forward to https://koboldai.org/discord where more people can help in a support channel. Otherwise you have to hope I see the notification in time.
57 replies