Hi all , iam used pagedcsvreader along with chromadb and when iam querying with similarity_top_k as 10 but none of the documents are relevant to query. Although iam directly specifying important keywords in query. What can I do to improve rag with CSV data
iam implementing a rag chatbot, but it's always answering from first retrieved documents only. If I can any other questions retrieved documents are same . It's not able to retrieve different set of documents
iam using document summary index for my context chat engine, now my chat chat engine only answers about summary or it can answer based on orginal document if some information is not captured in summary
I have multiple CSV files and data dictionary describing about each column . I want to use only open source llm and Rag to create conversational chatbot with memory. It should be able to perform aggregations also on my CSV data files
Hi how to use hosted llamacpp server in llamaindex for chat engine. Iam trying by importing OpenAiLike and passing model name and base_api still getting error
Hi All , do we have any llm server framework like vllm, open llm which can be run only on cpu and make use of multiple cpus in cluster like master and worker to serve multiple inferences parallel
@kapa.ai I am doing a rag chatbot with chat engine, how can I limit my output to strictly context and provide as less output as possible. Right now iam getting answer and also full follow up information about my question
Hi , I seen llmlingua and tried with llamaindex , also using llamacpp for loading llm. For each question the time taken to get prefix match hit is too high . Llm inference time although reduced, time taken to hit llm is so high that without llmlingua my chat engine is giving faster response. Any idea on this. Iam using only cpu