Hi π Is it possible to use llamaindex-cli rag with a local model? (preferably these that llama-cpp cli is able to run really fast with on an M1 Mac
https://huggingface.co/ggml-org). Looking at the source code, it looks like OpenAI is hardcoded into the CLI.