Find answers from the community

Home
Members
purple crow
p
purple crow
Offline, last seen 3 days ago
Joined December 26, 2024
Hello just starting to build a knowledge graph or long term memory for my LLM agents

evaluating llama index for the same.

Couple of quick questions -

  1. is there a s3 bucket loder in llama index
  2. i dont understand vector_store = ChromaVectorStore(chroma_collection=chroma_collection)
storage_context = StorageContext.from_defaults(vector_store=vector_store)
why do i need to send the vector store while load index and storage context both when storage context itself knows whats the vector store used ?
  1. do i need to store the embeddings at some blob storage as well, incase my vector store goes down i might need to recreate all of those ?
3 comments
W
p