Find answers from the community

Home
Members
Remo van de Sande
R
Remo van de Sande
Offline, last seen 3 months ago
Joined September 25, 2024
appreciated 💪 I think this topic already came to mind before in a talk by Lance from LangChain, summary:
(https://www.youtube.com/watch?v=UlmyyYQGhzc)
  • Context lengths for LLMs are increasing, raising questions about the necessity of external retrieval systems like RAG, especially when massive amounts of context can be fed directly into LLMs.
  • Greg Kamradt's Needle in A Haystack analysis tested LLMs' ability to retrieve specific facts from varying context lengths and placements within documents, revealing limitations in retrieval, particularly towards the start of longer documents.
  • RAG systems aim for multi-fact retrieval, requiring the retrieval of multiple facts from a context. Google's recent 100-needle retrieval demonstrates the need for efficient multi-needle retrieval for comprehensive understanding.
  • Retrieval from long contexts doesn't guarantee retrieval of multiple facts, especially with increasing context size and number of needles.
  • Cost of long-context tests can be managed effectively, with careful budgeting enabling meaningful research without significant financial strain.
Limitations for Longer Context:
  • no retrieval guaranties, multiple facts are not guaranteed to be retrieved, especially as the number of needles and context size increases.
    GPT4-0 tends to fail near the start of the document-size, less fails on bigger datasets.
  • Specific prompting is needed for larger contexts.
  • Performance degrades when the LLM is asked to reason about retrieved facts the longer the context.
  • Longer Context are pricey, and take longer to generate.
My takes: in the future there is less focus on indexing/chunking and more focus on improving retrieval while reducin hallucinations. DSPY could be interesting for this.
5 comments
m
R
has anyone been testing Llama_Extract ? we are looking for an affordable solution to extract meta-data from a large chunk of data using pre-defined schemas.
3 comments
R
b
just dropping this in here since it has been very helpful for us: https://superlinked.com/vector-db-comparison
3 comments
R
W
Thanks for letting me know as I already downgraded to 3.10 with no luck yet. Did you upgrade the import statements and are you also using DeepLake ?
1 comment
L
can someone help us with finding the updated module names/directories for:
Plain Text
from llama_index import download_loader
from llama_hub.github_repo import GithubRepositoryReader, GithubClient
from llama_index import VectorStoreIndex
from llama_index.vector_stores import DeepLakeVectorStore
from llama_index.storage.storage_context import StorageContext 
20 comments
L
R