Hi @Logan M, if I have some documents 50-100 pages, split in chunks of 1024, and I want to send full document to llm, will be ok to send all chunks or is better to create a new index with SummaryIndex.from_documents() and use this in query engine ? What option will have better results or will be the same ? or there is a better option ? Thanks
And if will have chunk overlap there can be any performance degradation sending all chunks versus, using SummaryIndex and send only one chunk with all document ?
I'm unsure whether sending all chunks would involve merging them into a larger chunk with the size of the context windows in my case 100k for Claude 2.1, or they're simply sent as a big batch without merging ?
using tree_summarize or compact response modes, it will already be stuffing the LLM input with as much retrieved text as possible (regardless of chunk size)