new_in_town
RRunPod
•Created by new_in_town on 10/8/2024 in #⚡|serverless
Is Quick Deploy (Serverless) possible for this RoBERTa model?
I want to run this model:
https://huggingface.co/MoritzLaurer/roberta-large-zeroshot-v2.0-c
as Serverless. And Quick Deploy (under https://www.runpod.io/console/serverless) shows multiple option, what should I choose? Should it be "Serverless vLLM" or something else? NOTE: this model is NOT a LLM. It is based on the RoBERTa architecture. And RoBERTa is not directly listed as a supported model architecture in vLLM
as Serverless. And Quick Deploy (under https://www.runpod.io/console/serverless) shows multiple option, what should I choose? Should it be "Serverless vLLM" or something else? NOTE: this model is NOT a LLM. It is based on the RoBERTa architecture. And RoBERTa is not directly listed as a supported model architecture in vLLM
2 replies
Suggest a template for this text classification model (small model from huggingface)
I want to do some (zero-shot) text classification with this model [1] or with something similar
(Size of the model: 711 MB "model.safetensors" file, 1.42 GB "model.onnx" file )
Now I see a LOT of pod templates...
Would you suggest a template for this model?
If possible: i would avoid writing Python code or use the minimal amount of Python.
[1] https://huggingface.co/MoritzLaurer/roberta-large-zeroshot-v2.0-c
3 replies