I am using
chat_engine = index.as_chat_engine(
chat_mode="context",
system_prompt="You are a chatbot with access to extra context information about dogs. Only answer questions if you can find the answer in the context, otherwise, kindly inform the user that you cannot find the answer.",
chat_history=custom_chat_history
)
All is working well and the LLM only answers questions about "dogs". My issue is that when I introduce the recent chat history, if someone asks a question about "cats", the LLM properly handles the quesiton the first try but any follow-up question including cats will be answered. Any recommendations for handling this situation?