well, at $0.002/1k tokens, 5K tokens is not bad 😅 Even with davinci at $0.02/1k, still could be worse
Llama index supports any local LLM, it's just up to you to pass the text to the model and return the new generated tokens
See this small example with FLAN ->
https://github.com/jerryjliu/llama_index/issues/544You mileage may vary though. Seems like every LLM needs slightly tweaked prompts. The default prompts are optimized for davinci and chatgpt