R
Railwayβ€’2mo ago
Mirsa

LLM Hosting on Railway

Hello πŸ‘‹ Can I host an LLM like ollama but able to respond to multiple requests on railway? I'm talking about deploying an LLM with an API capable of responding quickly to a medium number of requests.
Solution:
I'm no expert in LLMs but theres some templates like ollama and others in the marketplace so it's totally possible, now performance wise Railway doesn't offer GPUs so you would need a serverless GPU platform for it
Jump to solution
6 Replies
Percy
Percyβ€’2mo ago
Project ID: N/A
Mirsa
Mirsaβ€’2mo ago
N/A
Solution
Medim
Medimβ€’2mo ago
I'm no expert in LLMs but theres some templates like ollama and others in the marketplace so it's totally possible, now performance wise Railway doesn't offer GPUs so you would need a serverless GPU platform for it
Medim
Medimβ€’2mo ago
also if it's a CPU/RAM intensive process you may also consider upgrading to the Pro Plan
Mirsa
Mirsaβ€’2mo ago
Fine, thank's for you answer. I actually found those templates but i'm asking for a multi request, and ollama does not seem to be able to answer several requests simultaneously.
Medim
Medimβ€’2mo ago
Yeah, that's isn't related to Railway itself
Want results from more Discord servers?
Add your server