hi all, i am doing some project with locally installed llama 2 and following simple API interface:
{ "input": "how is weather in new york", "context":"new york is hot in these days" }
input the query and context should coming from the the vector DB. How i can get it integrate with existing lllamaindex library without change too much of my codes ? @WhiteFang_Jr
and in the normal api interaction with llm, i believe llamaindex will qurey the llm couples time for a re-fine. How llamaindex decide stop calling llm api and return the last answer back to user ?