Hi all,
I am using llama-index evaluation part to evaluate a RAG system, and I was wondering the theory behind those eval-questions. For example in context_relevancy, the prompt is
<"Your task is to evaluate if the retrieved context from the document sources are relevant to the query.\n"
"The evaluation should be performed in a step-by-step manner by answering the following questions:\n"
"1. Does the retrieved context match the subject matter of the user's query?\n"
"2. Can the retrieved context be used exclusively to provide a full answer to the user's query?\n">
https://github.com/run-llama/llama_index/blob/f5263896121721de1051ce58338a1e0ea6950ca7/llama-index-core/llama_index/core/evaluation/context_relevancy.pyDoes anyone know these questions are based on what principles?