Find answers from the community

Updated 2 years ago

I have a question about the GPT index

At a glance
I have a question about the GPT-index principle.
LLM has an upper limit amount of data, (token limit is 4096).
GPT-index compresses the external data into an efficient data structure.
Does the accuracy of each piece of information decrease as more external data is indexed and the compression ratio increases?

For example,
if I make ExternalDataA and ExternalDataB into one index and execute a query about ExternalDataA, will the accuracy be lower than when only ExternalDataA is indexed?
2
M
K
y
7 comments
I think it depends on the index type, but not sure
Nice to meet you. I apologize for the inconvenience.

My name is Koh and I am working on this issue with Sho.

We are both Japanese and we are very grateful to you for developing such a great open source.
@jerryjliu0

Please let me know if you know about the above problem.

We will look at the channel thread here and refer someone who may know more about
@jerryjliu0 @yourbuddyconner @Jeremy

If you think you can answer, we would appreciate it if you could answer when you are available.
Thanks, Mikko! @Mikko
If you're using a vector store, the natural result of doubling the index size is increasing the search space gpt_index has to search over. You'd have to start doing work of increasing similarity_top_k and/or improving queries such that the most relevant document is returned first.
@sho4360 @Koh to answer your question, we don't do "compression" of the data in the traditional sense of the word. Rather, we just store it in a format that we can pass to the LLM while obeying prompt size limitations. The simplest example is the vector store index as @yourbuddyconner mentioned, where we split your text into chunks under the hood, and when you call "query" on the data, we fetch the top-k text chunks and put it into the prompt, which we can then feed into GPT. Note that even if the top-k text chunks don't completely fit into one prompt, GPT Index can handle that for you, by calling the LLM repeatedly over sequential prompts.

Does this high-level idea make sense?
Thank you all for your answers.
I probably misunderstood about the prompt that GPT-index passes to LLM, but I understand it.
I will try adjusting the query such as similarity_top_k.
we do also allow you to create summaries easily, and store those per index though! which is kind of similar to compression
Add a reply
Sign up and join the conversation on Discord