Find answers from the community

Home
Members
hypoxia6424
h
hypoxia6424
Offline, last seen 3 months ago
Joined September 25, 2024
Hey team - Is Llamaindex safe for multiprocessing for both ingestion and querying? I'm interested in using it in a scaled application to replace my custom RAG setup, but haven't been able to find a recent answer to this question online or in the discord history.
6 comments
L
h
Hey team - Awesome work you're doing with Llamaindex! I'm keen to replace my existing RAG pipeline with it. Hoping / wondering if someone could help with two quick questions?

  1. Are there a caching mechanisms available out of the box for the OpenAIEmbedding and OpenAI classes? Development iteration speed is the top priority in my use case, and I've found it helpful to cache these results in my custom RAG pipeline.
  1. Has anyone developed and tuned an adaptive max_token mechanism for the semantic splitter? The semantic splitter is awesome but I keep running into chunks that are way too large when working with wikipedia data. Even dialing down the breakpoint percentile often doesn't work. Any tips and tricks would be greatly appreciated!
3 comments
h
L