Hey, I've built a chatbot using llama-index however I feel that they is a lot of latency to get an answer even if I only use a small vector index. Btw I'm using the free pinecone version. do you think it mainly do to this ?
Hi guys, I'm using the llama_chat_agent. However, each time he agent try to access a tool (index) this following error occurs --> AttributeError: 'ChatOpenAI' object has no attribute 'total_tokens_used' Does someone has faced this too ?
Hi, is it possible to modify the original prompt of the llama_chat_agent ?? keeping the same input and format just change some sentences in the prompt ?