The post is about a community member exploring LlamaIndex and having some questions about querying their indexes. The comments discuss how queries only search the indexed data, not the LLM's broader knowledge base. There is some uncertainty around how to constrain queries to the indexed corpus and how LlamaIndex interacts with other tools like LangChain. Community members share their experiences and insights, noting that LlamaIndex can help prevent LLM hallucinations and leverage the LLM's skills, but that chunking source texts and handling multiple relevant results are ongoing challenges.
Hello, I have been exploring llamaindex, managed to do some indexing and basic querying. I have a bit of a blindspot with my understanding, do queries only query my indexes?
I think sometimes I have a similar problem understanding what "querying your index" means @jerryjliu0 . I am assuming that the index is to provide context for LLMs to find answers within their "large knowledge base" rather than just from the documents we are indexing? This is the part I am not 100% sure about. If that makes sense?
In the prompt you can specify to not use prior knowledge ad to base the answer on your index…from what I experienced you use llama index to prevent llm to have hallucinations for specific answers and ofc to “add” knowledge taking advantage of llm skills
Just using the basic examples provided in the tutorials, I indexed a folder of technical documents. I was able to ask questions about things not covered by that corpus. So, I'm a unsure about my understanding, specifically about how to constrain the query to the corpus index..
ahh i see. yeah the way to understand a query is an initial "input prompt" that will be augmented with additional context from your data under llamaindex, and will give you the final result
Thanks for explaining that. As I delve deeper I am also finding a significant amount of overlap between llama index and langchain. Any advice on how you would combine the two? Currently my plan is to index with llama, store the results with pinecone then query with langchain?
I personally switched almost all to llama, and im losing all the news on langchain side. For now, querying with llama index gives be better results (maybe for the prompts and refining)
sometimes prior knowledge creeps in if your top hits contain irrelevant context (high embedding similarity but not answering the question). it also happens if you have multiple hits, after multiple rounds of trying to refine the answer the final answer gets screwed up. it's a combination of using prior knowledge and hallucination/confusion that you are seeing.