Oryza sativa
RRunPod
•Created by Oryza sativa on 3/25/2024 in #⚡|serverless
How to Run Text Generation Inference on Serverless?
but i just curious, it is using vllm right? so is runpod now support using TGI for deploying model in serverless?
https://github.com/huggingface/text-generation-inference
11 replies
RRunPod
•Created by Oryza sativa on 3/25/2024 in #⚡|serverless
How to Run Text Generation Inference on Serverless?
I am sorry, already get the response, I think becuse I hit the endpoint but my endpoint still on initializing status, isn ot ready yet. and finally get the response with this. thank you @ashleyk
11 replies
RRunPod
•Created by Oryza sativa on 3/25/2024 in #⚡|serverless
How to Run Text Generation Inference on Serverless?
11 replies
RRunPod
•Created by Oryza sativa on 3/25/2024 in #⚡|serverless
How to Run Text Generation Inference on Serverless?
is it also support text generation inference?
11 replies