Flash attention 2 is supported in streamlit free app deployment?

I am building an app where I need repo id: suno/bark-small, this model supports Flash Attention 2 to decrease the inference time. The Flash Attention 2 only supports the latest GPU

More details: suno/bark-small · Hugging Face

:rocket: Deployment llm discussion

This topic was automatically closed 180 days after the last reply. New replies are no longer allowed.