QQ, I turned on logging when using a Huggingface embedding and I see it always connects to Huggingface, even if I already have the model cached. Looks like it grabs the tokenizer_config and config.json each time, even though I see those files in the cache folder. Any way to tell it to stop?
from llama_index.embeddings.huggingface import HuggingFaceEmbedding
# loads BAAI/bge-small-en
# embed_model = HuggingFaceEmbedding()
# loads BAAI/bge-small-en-v1.5
embed_model = HuggingFaceEmbedding(model_name="BAAI/bge-small-en-v1.5")