Flash attention 2 is supported in streamlit free app deployment?

I am building an app where I need repo id: suno/bark-small, this model supports Flash Attention 2 to decrease the inference time. The Flash Attention 2 only supports the latest GPU

More details: suno/bark-small 路 Hugging Face

:rocket: Deployment llm discussion