I am trying to build "another" RAG system, using tools, with data stored in neo4j, chroma and summary index. the issue m facing is, the response time is around 25 min. the query is a bit complex too, but is it because of my local system? I am using OpenAI gpt-4-1106-preview model. I am running this locally on my macbook M1 chip with 8GB Ram. can anyone kindly guide me what i am doing wrong here?
Hi Liqvid. GPT-4 is known to be slow (it's so popular that it's almost always jammed up). Would your use case work with GPT-3.5? (Or better, use a locally-served LLM so that you won't have to pay a dime?)