Anyone have a solution in their RAG app for smart date parsing? E. G. To filter docs in a date range based on the query. I’m aware of things like spacy which can identity date entities, but then there is a missing intermediate step to go from something like “this year” to an actual date/time for a query filter.
Ok next issue 😊 trying to get streaming to work. Im using a langchain llm (HuggingFaceTextGenInference) and the streaming works from my inference endpoint. However, when using it with llama_index I get error “LLM must support streaming”
Having an issue with json mode.. I’m getting thousands of new lines from my request with the actual response somewhere in the middle. Anyone seen this?
Having an issue with azure openai.. but only when running in a thread executor. Anyone else deal with this? Is the azureopenai object no longer thread safe? Was working before I upgraded llama index and openai packages
Having trouble with azure. I have a key and everything set up, I set my environment variables, but it still says you must set OPENAI_API_BASE. I’ve confirmed it’s correct and available in the environment.
Hey all, is it possible to limit the chunk size in the node parser to be sentences? I have much better results with my data using sentence embedding vs embedding larger chunks. My current process is to use spacy to identify the sentences semantically and then pass them to my embedding model. This is critical for the types of problems I’m trying to solve. Also, I wrote an api around this whole process, would be nice to synthesize it with the node parser somehow..