Find answers from the community

s
F
Y
a
P
Updated 7 months ago

QQ on vector stores. I've been playing

QQ on vector stores. I've been playing first with the native json storage method, then chromadb, and lancedb. Using the same data set that generated a 50 mb json file, all these methods seem to take a minute to load the index, and they don't max our the memory or CPU on the machine during load.

Do these "serverless" methods all just take a long time to load at first? Is the entire dataset loaded into memory?

If I move to postgres or similar I assume the "index load" will be much faster?
L
P
v
10 comments
yea others will be much faster imo, if it doesn't have to read from disk constantly onload
like running qdrant in a docker container, using weaviate or pinecone
Is this due to how llama index interacts with these solutions versus a real server based one? When I see these serverless solutions making claims of data sets with millions of columns being performant, seems counter intuitive from this experience...

Anyways I'll setup postgres and compare
Server-based APIs over all IMO
I've been real happy with Qdrant, super easy to deploy if you are into self-hosting, very fast
Nice, thanks for that. Yeah trying the full self-host method for now. Setup postgres last night but wasn't able to test too much. Qdrant is on deck next.
Are you using Kubernetes? (I have deep experience with it, so I am for this project I'm working on, would recommend the Zalando Postgres Operator)
Not yet, just deploying everything locally on a Windows computer. Postgres + pgvector is rippin' so far.
I recently got my prototype into a production like environment. I'm a big proponent of "The Tracer Bullet" concept in software
https://growsmethod.com/practices/TracerBullets.html
I had to change a bunch of things to get my code to work both locally and in the cloud. (mostly non-ai stuff, but I also switched to using an API for embeddings)
Add a reply
Sign up and join the conversation on Discord