Hi How to response stream output when using agent that built via initialize_agent(), I can't do it following your notebook which wrote for as_chat_engine(). Because the function agent.run() only return string type.
Once I assignment graph.index_struct.summary a value ,next time I load this graph from disk .This value will be none. Is there a bug of save? Or does this field have any effect?
hello bros, Can I pass two prompts by one query? How to do it? index = GPTListIndex(documents) response = index.query(prompt, response_mode="tree_summarize")
Hi. I got following error when chat to agent: openai.error.InvalidRequestError: This model's maximum context length is 4097 tokens. However, you requested 4918 tokens (3894 in the messages, 1024 in the completion). Please reduce the length of the messages or completion.
@Logan M I had create an agent using create_llama_chat_agent,based on ComposableGraph,and I add a GPTSimpleVectorIndex with document to ComposableGraph.Then start chat to this agent.But I found this the agent's answer that were rarely relevant to my document.I want know how to improve the hit rate of index against my documents?