Noob question about working with large (230 mb) index? Hi, I used GPTSimpleVectorIndex to create an index of several hundred pdf files. When I try to query the index, I get answers but they are kind of inconsistent -- makes me think(?) that query is only grabbing some information each time but then running out of memory?
Right now I'm using the index.query as described in the "getting started" tutorial. Very new at this and so would appreciate pointers?
I reviewed recent videos on youtube that suggested
- using pinecone to increase memory size (but does that improve processing?)
- video that suggested progressive summarization (query applies to one chunk at time, then all answers are concatenated together and then fed back into GPT, which provides summary of the concatenation of all the answers)
- video that suggests using langchain.
Any pointers appreciated. Thanks!