however worth noting that gpt_index retrieves from the vector store and uses an LLM call to answer your query so if you want to be completely "offline" you'll have to work with the store directly (not a huge deal)
Although, now that I think about it, I think inserting into the SimpleVectorStore is an offline process so there might be a local model doing it in gpt_index...
Jerry would know more he'll probably chime in eventually
Also: how does gpt-index relate to model fine-tuning?
Kind of orthogonal, gpt_index is used so one doesnt need to fine-tune a model over a dataset. It facilitates retrieval of a subset of a document for use in a prompt.