Hello @Logan M and everybody I want my llama-index RAG chatbot to remember past question for context, can I follow the format openai has for chat completion? here message will go to llama-index bot instead of this?
response = client.chat.completions.create( model="gpt-3.5-turbo", messages=[ {"role": "system", "content": "You are a helpful assistant."}, {"role": "user", "content": "Who won the world series in 2020?"}, {"role": "assistant", "content": "The Los Angeles Dodgers won the World Series in 2020."}, {"role": "user", "content": "Where was it played?"} ])
also, how to manage if message get close to context length for give openai model? is there any better way to make model remember past questions?
Hello @WhiteFang_Jr what you're saying is correct but I am using multi llm architecture so how can I pass last/final output to this first llm for it to remember past question? is there any way? really appreciate your help.