Find answers from the community

Home
Members
cmosguy
c
cmosguy
Offline, last seen 2 hours ago
Joined September 25, 2024
@kapa.ai I am attempting to use the multiagent workflow but it flakes out and does not hand off to all the agents as it is supposed to
25 comments
c
L
k
@kapa.ai and @Logan M lets say I am using multi-agent workflows. Is it possible to have some type of chat memory stored and injected in the next round of multie agent workflow?
7 comments
k
c
c
cmosguy
·

Research

@Logan M and @jerryjliu0 I’m sure you geniuses at llama index have figured out a way to replicate this in your framework and python: https://jina.ai/news/a-practical-guide-to-implementing-deepsearch-deepresearch/
2 comments
L
c
@kapa.ai please show me an example of how to make a customqueryengine with reranker an dnode post processors
27 comments
k
c
@kapa.ai how can i setup the Vector store as a remote endpoint - I am trying to debug a class and I do not want to keep reloading the vector store over and over, I just want this to be a separate process that I can reach via a socket or api somehow, is there something in llama index api that can do this out of the box?
19 comments
k
c
@kapa.ai I keep seeing httpx in the warning output of the posts going to azure open ai is the something in llama index that keeps showing this only in my Jupyter notebooks
6 comments
c
k
c
cmosguy
·

O3

@Logan M how did I use o1 or o3 as an agent in the workflow agent system? Do you guys have an example?
8 comments
L
c
@kapa.ai in the agent workflow. It looks like I have this timer. How do I know is an agent time out? Is there something the In the handler for the workflow that indicates a time out was reached?
2 comments
k
Has anyone started to build an MCP service to llama index yet?

https://modelcontextprotocol.io/docs/first-server/python
13 comments
b
L
0
@kapa.ai I want to take some text and reformat the text using an LLM, what is the best way to do this, should I use acomplete method?
7 comments
k
c
@kapa.ai lets say i am using the AgentWorkflow and I want a planning agent to use OpenAI o1 models that do not allow for system prompts and cannot do function calling, how would you recommend building this as an agent system to tie in the AgentWorkflow?
3 comments
k
@Logan M what I am asking is how does one think through scenario where you have multiple query sources. In this example you may have search_web search_internal_db search_custom_db , lets say you want to the tool to search all three, so does it search all 3 of them in parallel or should a special retriever be created instead to manage the different retriever endpoints, does that make sense?
5 comments
L
c
@Logan M @kapa.ai something strange is happenig when I run my documents through the pipeline I get this error:
Plain Text
    nodes = await pipeline.arun(documents=processed_documents)
    pipeline.persist(persist_dir=pipeline_cache)


Traceback (most recent call last):
File "<string>", line 1, in <module>
File "/slowfs/amsgcn4/home/kleinada/ds_work/snps-edag-cc-dataset-gen/venv/lib/python3.11/site-packages/llama_index/core/ingestion/pipeline.py", line 319, in persist
self.docstore.persist(docstore_path, fs=fs)
File "/slowfs/amsgcn4/home/kleinada/ds_work/snps-edag-cc-dataset-gen/venv/lib/python3.11/site-packages/llama_index/core/storage/docstore/simple_docstore.py", line 84, in persist
self._kvstore.persist(persist_path, fs=fs)
File "/slowfs/amsgcn4/home/kleinada/ds_work/snps-edag-cc-dataset-gen/venv/lib/python3.11/site-packages/llama_index/core/storage/kvstore/simple_kvstore.py", line 88, in persist
f.write(json.dumps(self._data))
^^^^^^^^^^^^^^^^^^^^^^
File "/u/kleinada/.pyenv/versions/3.11.9/lib/python3.11/json/init.py", line 231, in dumps
return _default_encoder.encode(obj)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/u/kleinada/.pyenv/versions/3.11.9/lib/python3.11/json/encoder.py", line 200, in encode
chunks = self.iterencode(o, _one_shot=True)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/u/kleinada/.pyenv/versions/3.11.9/lib/python3.11/json/encoder.py", line 258, in iterencode
return _iterencode(o, 0)
^^^^^^^^^^^^^^^^^
File "/u/kleinada/.pyenv/versions/3.11.9/lib/python3.11/json/encoder.py", line 180, in default
raise TypeError(f'Object of type {o.class.name} '
TypeError: Object of type PosixPath is not JSON serializable
12 comments
k
c
L
@kapa.ai I do not understand when to use the RetrieverQueryEngine explain to me why I would use this
3 comments
k
@kapa.ai I have a markdown file with an image embedded in it. Is there a reader that creates both the document and image document with the images being described with a multi modal model?
6 comments
k
c
@kapa.ai I was trying to store a list in a value called "assciated_images" key in the metadata, but i got this error: ValueError: Value for metadata associated_images must be one of (str, int, float, None)
12 comments
k
c
@kapa.ai I am trying to embed my directory using open clip, but how do I tell what is the text associated with the document that it is embedded with?
41 comments
k
c
@kapa.ai I want to create custom response syntehsizer because i have metadata that I want to inject into the context window to help with answering the prompt how do I do this? I want to add the metadata to what is synthesized
11 comments
k
c
@kapa.ai what is the difference between TokenSplitter and SentenceSplitter, which items are the best to use?
28 comments
k
c
@kapa.ai when I load nodes into a vector store how do I avoid duplication? how do I just refresh the nodes?
23 comments
L
k
c
@kapa.ai how do I get all the nodes from the vector store I store all the nodes via a pipeline?
2 comments
k
c
cmosguy
·

Storage

I’m seeing a significant slow down when I retrieve a storage connect from disk - is this there simmering I can do to figure out what is doing in here?
5 comments
L
c
@kapa.ai I have microsoft word documents, how do I extract the document content into a markdown and include the images as well?
14 comments
W
k
c
Hey @Logan M how are you? I am trying to use the workflow example to generate an example sub questions, then go through and use react to answer the subquestions form here: https://docs.llamaindex.ai/en/stable/examples/workflow/sub_question_query_engine/

The issue is when I get to the point in the subquestion routine:

agent = ReActAgent.from_tools(
await ctx.get("tools"), llm=llm_4o_2, verbose=False, max_iterations=5
)
response = agent.chat(ev.question)

There are some subquestion quries where it fails with :

Error code: 400 - {'error': {'message': "This model's maximum context length is 128000 tokens. However, your messages resulted in 129643 tokens. Please reduce the length of the messages.", 'type': 'invalid_request_error', 'param': 'messages', 'code': 'context_length_exceeded'}}

I really do not understand how to control this. BTW, the tools is a lit of retriever tools, that was supposed to have node_postprocessor reranker to titrate down the nodes. but i do keep hitting this error regardless.
6 comments
c
L
@kapa.ai I am trying to use an LLM azure open ai model, the chat works but the .complete Keeps asking for api key
19 comments
k
c