Currently, I write the code like this because I also want to filter specify data in .csv file but I am not sure it is practical ? pdf_nodes = parser.get_nodes_from_documents(list_of_documents) index = VectorStoreIndex(list_of_documents)
@Leonardo Oliva This is my full code but I am sure it can be applied your method? df_Perf = pd.read_csv('/content/gdrive/My Drive/llama_index/VW_T_LEAVE_DETAIL.csv') list_of_documents = [] for i in range(len(df_Perf)): list_of_documents.append(Document().encode) pdf_nodes = parser.get_nodes_from_documents(list_of_documents)
@Leonardo Oliva Thanks you, however, the result is incorrect answer when I asked some information in .csv file. Can you give me the advice what I should check it?
@Leonardo Oliva How are you Splitting your text? =>> SentenceSplitter (chunk_size = 512) What's your retriveal strategy? (I don't understand this question), Which LLM are you using? =>> gpt-3.5-turbo, model ='text-embedding-3-small, max_tokens=512