How to sharing my Streamlit app with large-scale ML models to multiple users

I’d like to run my Streamlit app for multiple users so that some users can access it at the same time.

The app includes large-scale pre-trained language models and it exceeds the limit indicated in Deploy an app — Streamlit 0.84.0 documentation, hence I plan to run the app in another way.

I’ve checked the discussion in Does streamlit is running on a single-threaded development server by default or not?, and still wonder my app using such a large model with @st.cache(suppress_st_warning=True, allow_output_mutation=True) can be run in multi-threads.

Does the cached deep learning model support access from multiple threads?

Thank you in advance.