Find answers from the community

Updated 11 months ago

Hey team - Awesome work you're doing

Hey team - Awesome work you're doing with Llamaindex! I'm keen to replace my existing RAG pipeline with it. Hoping / wondering if someone could help with two quick questions?

  1. Are there a caching mechanisms available out of the box for the OpenAIEmbedding and OpenAI classes? Development iteration speed is the top priority in my use case, and I've found it helpful to cache these results in my custom RAG pipeline.
  1. Has anyone developed and tuned an adaptive max_token mechanism for the semantic splitter? The semantic splitter is awesome but I keep running into chunks that are way too large when working with wikipedia data. Even dialing down the breakpoint percentile often doesn't work. Any tips and tricks would be greatly appreciated!
L
h
3 comments
  1. Hmm, no adaptive max token mechanism. But you could take the resulting chunks and pass them through a secondary TokenTextSplitter or SentenceSplitter to catch chunks that are too big
Ah thanks @Logan M πŸ™‚
Add a reply
Sign up and join the conversation on Discord