Yes I guess, you mentioned it already. Thanks for that. But actually that "missing prompting documentation" should somehow be addresses. Basically three issues:
Is the tool description used as prompt and best practices what to put there
I try to icecream agent.get_prompts, this is always empty? Is this correct?
Sometimes I get "Could not parse output: Thought: ..... " what might be the cause of this?
I have one more question and would be very happy if you could answer it: So, my understanding is like this: Let's say i have a simple document store with one vectorized document which are in fact embeddings. Now my query is tokenized and this is converted into embeddings. Now there is (my understanding) a "cosine similarity" or something similar search for some "hit" in the vector store. Now the important question: How is that "hit", means the "matching" embedddings from the vector store fed into the llm?
The top-k matching results (calculated with cosine similarity or similar) are retrieved, and given to a response synthesizer. That response synthesizer puts the retrieved nodes + user query into a prompt template, and asks the LLM