I have already built a solution to create custom AI chatbots from own data which uses gpt-3.5-turbo and RAG to fetch the context using openai embedding and cosine similarity match.
I am facing issues when user asks some follow up question on the previous question, in this case doing RAG for just the question is not sufficent as it will not fetch the required context.
I already tried question rephrasing based on conversation history but it makes responses slow and difficult to decide which one to rephrase and which one to not.
I went through the source code of chat llama_index and couldn't find anything different, follow up questions are not handled is what I observed.
How RAG will work for this scenario? Q: what is apple watch? A: bla bla Q: what are it's features?
I have tried chatbase.co and it works pretty well don't know how they do it. Can someone please help me on this
I have already built a solution to create custom AI chatbots from own data which uses gpt-3.5-turbo and RAG to fetch the context using openai embedding and cosine similarity match.
I am facing issues when user asks some follow up question on the previous question, in this case doing RAG for just the question is not sufficent as it will not fetch the required context.
I already tried question rephrasing based on conversation history but it makes responses slow and difficult to decide which one to rephrase and which one to not.
I went through the source code of chat llama_index and couldn't find anything different, follow up questions are not handled is what I observed. Can someone please help me on this
Hi Everyone, is there a tool or abstraction in llama index with which i could get details scraped from the internet for example:
"Im a buyer at Walmart. Im looking for suppliers for: Carton boxes with window for Pizzas in North America: The US, Mexico and Canada. I need a list of 200 suppliers. Just the suppliers and their website."
Hii, I am looking for a notebook which has end to end RAG pipeline suitable for production use case, like it can handle tables from pdf, also handle summary questions and basic question answering. Please share links if there exist such master notebook.
Hi Everyone, Does llama index use semantic router to route queries or it still uses LLM generations to make tool-use decisions, semantic router is lot faster i guess, wondering if llama index has integration to it https://github.com/aurelio-labs/semantic-router
Hi, is there a way to use existing namespace in pinecone as query engine in llama index, I ingested data some other way and now i want to connect it to llama index
Hi Everyone, is there a way to do text chunking in llama_index dynamically, for example i want the chunk overlap to be 2 sentences from last chunk until some threshold is hit.
Hi Everyone, how do i make sure my RAG chatbot doesn't answer out of context questions like "write poem on moon", i'm using gpt-3.5-turbo i have tried providing instruction in system and user prompt but doesn't work everytime, chatbots starts to write poem when it should be saying "Sorry, I don't have information about it."