codyman4488
codyman4488
RRunPod
Created by codyman4488 on 3/4/2025 in #⚡|serverless
how to run a quantized model on server less? I'd like to run the 4/8 bit version of this model:
5 replies