Find answers from the community

Updated 3 months ago

Questions

needed some explanation on a few concepts in the llamaindex documentation :-

  1. https://gpt-index.readthedocs.io/en/latest/end_to_end_tutorials/dev_practices/production_rag.html - what do you mean by "Decoupling chunks used for retrieval vs. chunks used for synthesis" - the synthesis happens on the chunks retrieved only right? How can they be decoupled?
  1. https://gpt-index.readthedocs.io/en/latest/examples/retrievers/auto_vs_recursive_retriever.html - How is VectorStoreInfo working here? I can see we have set metadata for every node at the top of the article. How is it getting connected to the vector_store_info property?
L
r
4 comments
  1. Decoupling just means you take the retrieved nodes and do something to them before sending them to the LLM (I.e. a node postprocessor). The sentence window retrieval is a good example of this. Retrieve sentences, replace with the surrounding context
  1. Vector store info is being used in the auto retriever. It is used to format a prompt that helps the LLM write query settings
How is it helping in refined answers to queries as mentioned in the doc? I am actually trying to understand what's happening in the backend a little more. Deeper explanation than - " It is used to format a prompt that helps the LLM write query settings" πŸ™‚ πŸ₯Ί
Add a reply
Sign up and join the conversation on Discord