Find answers from the community

Updated 10 months ago

hi guys i have a question on one of the

hi guys i have a question on one of the latest video/documentation from llama_index on LLMs for Advanced Question-Answering over Tabular/CSV/SQL Data (Building Advanced RAG, Part 2)

Jerry Liu is indexing all his files first (he got tons of CSV files) but i was trying to do the same on my SQL, one of my table only has 27,000 rows (the other one is like 1 milions rows) and even the small table it tooks ages to index. i know his doing that to then use get_table_context_and_rows_str to give the AI some relevant rows

how can i do it ? is it because im supposed to save my SQL table data into CSv? is that why its taking so long ?
L
M
4 comments
Its taking a long time to embed I'm guessing?
well, i just saw your reply on youtube, but i was wondering in the video of Jerry, it was fast, was it becasue each of his "table" was very small ?

i dont use CSV, i basicall do select * from {mytable} but the table is 27,000 rows thats the smallest table, the JSON file went up to 800mb, imagine the table of sales of 1 million rows, it does not make any sense

but i do love what hes doing but i wonder how i can make it happen (he was solving the problem to give some rows context for the "B.I.G" dot problem"
Most likely its because it was a small table yea

For bigger tables, bumping the batch size, as well as using a vector db integration (qdrant, chroma, etc.) will probably help a bit πŸ‘€
okai i will have to see about that....
Add a reply
Sign up and join the conversation on Discord