lawtj
lawtj
RRunPod
Created by lawtj on 9/16/2024 in #⚡|serverless
Error requiring "flash_attn"
I'm trying to run MiniCPM-V which according to docs supports VLLM (https://github.com/OpenBMB/MiniCPM-V/tree/main?tab=readme-ov-file#inference-with-vllm), but on run I'm getting ImportError: This modeling file requires the following packages that were not found in your environment: flash_attn. Run pip install flash_attn Any help on how to overcome this error? I was trying to use the webUI to configure serverless.
5 replies