Log in
Log into community
Find answers from the community
View all posts
Related posts
Was this helpful?
π
π
π
Powered by
Hall
Active
Updated 4 weeks ago
0
Follow
Cached Augmented Generation (CAG) with Gemini or other LLMs integrated with LlamaIndex
Cached Augmented Generation (CAG) with Gemini or other LLMs integrated with LlamaIndex
Active
0
Follow
A
Alwin
last month
Β·
Hi everyone,
Is there any implementation of Cached Augmented Generation (CAG) with Gemini or other LLMs integrated with LlamaIndex?
L
A
5 comments
Share
Open in Discord
L
Logan M
last month
correct me if I'm wrong, but doesn't CAG require direct model access (i.e. with pytorch)? I don't think you can implement this over an API
L
Logan M
edited last month
You can see an example here:
https://github.com/hhhuang/CAG/blob/main/kvcache.py
A
Alwin
4 weeks ago
Thanks for sharing that with me.
I'm unclear about direct model access too, which is why I'm looking into how it might work with Gemini. It might not be doable!
A
Alwin
4 weeks ago
As I know, Gemini has the longest context window among the LLMs, so CAG is really only meaningful with Gemini!
L
Logan M
4 weeks ago
There's lots of local models with large context windows too. But of course not as big as gemini
Add a reply
Sign up and join the conversation on Discord
Join on Discord