The community member is working on a project using the locally installed Llama 2 model and a simple API interface. They want to integrate this with the existing LlamaIndex library without changing too much of their code. The comments suggest that the best approach is to implement the LLM class, providing an example from the LlamaIndex documentation. The community members further discuss how to set up the API access points and how LlamaIndex decides when to stop calling the LLM API and return the final answer to the user. However, there is no explicitly marked answer in the provided information.
hi all, i am doing some project with locally installed llama 2 and following simple API interface:
{ "input": "how is weather in new york", "context":"new york is hot in these days" }
input the query and context should coming from the the vector DB. How i can get it integrate with existing lllamaindex library without change too much of my codes ? @WhiteFang_Jr
and in the normal api interaction with llm, i believe llamaindex will qurey the llm couples time for a re-fine. How llamaindex decide stop calling llm api and return the last answer back to user ?