Hello, I have an app which runs fine locally, gives what I want using an openAI LLM.
When I deploy this to google cloud, I see an open AI token error saying,
“InvalidRequestError: This model’s maximum context length is 4097 tokens, however you requested 519913 tokens (519657 in your prompt; 256 for the completion). Please reduce your prompt; or completion length.”
My code only returns a list of 10 numbers.
I am not sure where this is coming from, could this be because streamlit couldn’t end a user session or something?