I have built an RAG chatbot using "meta-llama/Llama-2-7b-chat-hf" model to let model act as a document question answering chatbot and below are the model params
I have used VectorStoreIndex to index my documents. When I query my model it is able to give answers related to the document accurately( Which is a good thing), but it is also answering questions that are out of my documents, I want to limit it to answer questions only related to my docs, How can I do that?
Have you included that instruction in the prompt? But in general with the smaller and less capable models they will still be prone to hallucinating in that situation
Yeah I included in the instruction prompt to strictly stick to the source documents, it got better than before but still it is answering some things out of document, should I give a more enhanced instruction prompt??