Thats correct, generated_dataset_from_nodes generates questions specific to each nodes. The generated dataset a) tests semantic retrieval b) tests llm generation
Questions like "Are there any invoices about graphics cards?" don't really make sense for vector retrieval? But in any case, to generate questions like these, you'd need context of the entire dataset (a summary?), and then you could just prompt an llm directly to generate questions
Thank you for the quick response! I am trying to test the downstream performance of my rag, so i was thinking i need a mix of questions. Some that represent the all the documents or a subset. I feel like a summary will remove critical information. Is there a better way to do this?