My set up is like this. User sends a query to the RAG based chatbot (condense plus context one). There's a similarity cut off that doesn't pass irrelevant context to LLM. Is there any way, if the number of documents retrieved is zero, I can skip this call to LLM and return a generic response. @Logan M
@WhiteFang_Jr You can see the context is empty. If I use only condense question then it is giving empty response. But that is not the case for condense plus context. Any ideas or suggesstions?