I’d like to run my Streamlit app for multiple users so that some users can access it at the same time.
The app includes large-scale pre-trained language models and it exceeds the limit indicated in Deploy an app — Streamlit 0.84.0 documentation, hence I plan to run the app in another way.
I’ve checked the discussion in Does streamlit is running on a single-threaded development server by default or not?, and still wonder my app using such a large model with
@st.cache(suppress_st_warning=True, allow_output_mutation=True) can be run in multi-threads.
Does the cached deep learning model support access from multiple threads?
Thank you in advance.