Log in
Log into community
Find answers from the community
View all posts
Related posts
Did this answer your question?
π
π
π
Powered by
Hall
Inactive
Updated 10 months ago
0
Follow
I assume this comes up time and time
I assume this comes up time and time
Inactive
0
Follow
c
chirag
10 months ago
Β·
I assume this comes up time and time again when models release new context windows, but curious where RAG, vector dbs, etc comes into play with these attention focused models?
L
c
5 comments
Share
Open in Discord
L
Logan M
10 months ago
You are referring to models with huge context windows right? (i.e. Gemini 1M context window)
I think RAG will always play a role. Its similar to how computers have L1, L2, L3 cache, seperate from RAM, separate from harddrives
Smaller sizes mean faster speeds, less costs
L
Logan M
10 months ago
RAG helps the input to an LLM remain small (but accurate)
c
chirag
10 months ago
sure, cost and speed are relative now, but time has shown that these things will converge
L
Logan M
10 months ago
Has the speed of an L3 cache converged with RAM yet? Or Hardrives? π
Idk, there is always going to be a cost to input sizes. People will make them more efficient, but that just means people will also make them bigger
L
Logan M
10 months ago
This article articulates some points as well
https://vectorize.io/2024/02/16/rag-is-dead-long-live-rag/
Add a reply
Sign up and join the conversation on Discord
Join on Discord