As to your other point, assuming your loaded documents have a consistent doc_id, you can attach a docstore + vector store to the ingestion pipeline (redis, mongodb, firestore, postgres) and then manage upserts that way
Just browing your repo! Amazing work man. Only thing I need to really take care of is pulling rate to meet concurrency limits (HF inference server sets it to 512 conc requests), do you have any feedback on this end?
Thanks again! One last question, i swear. If my documents are continuously addeded to the Vector DB as they’re processed in the ingestion pipeline, is there any way to update/refresh on a live basis also the associated index? This is needed to smoothly allow RAG over new documents too, but Im not able to find anything similar from llamaindex docs, it seems like indexing is kinda static concept atm