I aim to develop a Streamlit application featuring a ChatGPT-like user interface for question-answering. The app will handle hundreds of pre-configured PDFs using the quantized version of Llama 2. However, a challenge is that each time a question is asked, it creates the embedding and loads the model.
This topic was automatically closed 180 days after the last reply. New replies are no longer allowed.