I really liked this stuff you guys did on putting RAG in production. The ideas seem promising, but I still find that they fail on various edge cases, particularly in technical documents. For example the windowing idea, where you index one sentence at a time, but then the LLM sees a larger portion of the document. I end up with a lot of garbage in my index :D. Also, PDFs with tables π. Are you doing some more work / discussion around this?
https://docs.llamaindex.ai/en/stable/end_to_end_tutorials/dev_practices/production_rag.html