I'm making the course of Llama index on deeplearning.ai and my question is if it is possible to make a RAG with an Open source LLM like dolphin mistral. All the examples I can find so far is always using OpenAI but no information to make RAGs with LlamaIndex with other models. Appreciate any link or suggestion that helps me clarify this. Thanks!
Yea, llama-index has various ways to sneak around these limitations.
For example, the default response mode of a query engine is "compact and refine" -- pack each LLM call with as much text as possible, and refine an answer to a query across 1+ LLM calls.