Hello there, I have a vanilla rag & reranker pipeline set up. my rag application is suppose to retrieve all line items from an invoice with an unknown number of line items. Could range from 1 line item to tens or hundreds of line items.
In my prompt, I have explicitly stated to retrieve all line items from my index, however, at all times the retrieval is always incomplete. Suppose the retrieval fits into the LLM response context window, what can I do to ensure it always return to me all line items in my index?
im running my pipeline on 8 page document and all pages contain tables. each table consisting of around 2 to 7 line items (which could be different if the document comes from another source)
hey @Logan M just wanna do a quick check. I also have a FlagEmbeddingRanker with a top_n=2 included in my query engine. It does also affect the number of chunks returned in a query right?
Hi @Logan M I'm still having trouble with my application. I have created a question issue in the LlamaIndex github repo with sample docs and reproducible codes. Can you or perhaps anybody in your team help me out? https://github.com/run-llama/llama_index/issues/15075