Is there a "dumb" model that is super fast and just finds matching vectors from a vectorstore database. like i give a prompt about something i want and it only does the query part without the inference.
I hate the overly dependence on ollama and openai stuff in the docs these days. half a year ago everything could be setup locally, now llamaindex docs is bloated with this 3rd party software and always download embeddings and every little thing online