Find answers from the community

Updated last year

Logan M When I m using Multi hop Agent

@Logan M When I'm using Multi-hop Agent calls, there's increased response times for queries especially when calling multiple tools across agents. Is there a caching solution that I can use to help reduce response times?
L
V
8 comments
We've been meaning to integrate something like gpt cache, but we haven't gotten around to it yet lol
I agree it would probably be helpful
What about implementations like Langchain's InMemoryCache or something
I think you'd have to use a langchain LLM class (which then should have caching)
And that should be compatible with llama index 🫰
Well Langchain LLMs aren't compatible with OpenAI Agents iirc
What I meant was more so porting what they are doing for caching, natively to llama index somehow
Add a reply
Sign up and join the conversation on Discord