could you share more details? how dod you try it, and what you would like to achieve
hi Emanuel, I'am trying to design a way to submit prompts like:
Prompt: list the top 10 cited <elements such as words/topics/etc> in the provided documents
for instance asking for top 10 used words in newspaper articles
I think this kind of retrieval of information should be somehow supported on a database level, not only LLM
cause obviously, the plain model will search for the most similar chunks in its context, without being able to retrieve what I am looking for
idk probably this is a weird and dumb question, but I struggled to find something useful in that sense diving in documentation
also in chromadb/pgvector/pinecone documentation at query-level
all the efforts seem to be focused on similarity retrieval only
Yea this is less about retrieval, and more about prompt engineering no?
what kind of prompt engineering could you setup to ask for frequency of words?
i struggle to have ideas in that sense
thanks, I will dive in the article you provided!
regarding the prompt you suggested, kinda black boxed. Isn't it? I mean, what is going on behind it?
I mean, LLMs are black boxes to begin with π
I don't think you'd want to hear a technical explanation of how LLMs and transformers work.
Basically, give them a prompt and some context, and it hopefully follows your instructions if the LLM is capable enough
but I mean, isn't one of the purposes of frameworks like Llamaindex to provide developers tools to design less-blackboxed apps with LLMs?
that's my point, but you are totally right
I just find vectordbs very limited in their query abilities at the moment, but probably it's due to their relative youth
by the way (last question I swear!), if I provide my application with thousands of articles from newspapers asking for what you said List the most common topics in the provided contex
, don't you think my costs will explode and results poor without proper llamaindex tuning??
exactly, processing every single piece of text is crazy. I was thinking about some approaches to handle this but using smaller local task-related-models it definitely a thing!
In your opinion these kind of models integrations should be used during ingestion (i.e. metadata extraction) or retrieval (i.e. augmentation strategies)?
I know, I have a lot of questions but I just find it too incredibile to just sit there and not ask ahaha
no worries! I think it depends on your use case as to where to use them
Using during ingestion for metadata extraction will help with retrieval
Using them during retrieval (like a node-postprocessor) can help modify retrieved chunks and change what gets sent to the LLM
can you help me find relevant pieces of llamaindex's documentation useful to understand how to develop such kinda-of data agents provided with small local models?
thanks! yea I mean, I was wondering how can I equip agents and/or other entities with, for instance, the models you linked here. Did you mention them not to be used into llamaindex envs but as outside modules?
kinda creating a tool with it for a data-agent (that's where my question about data agents came out)
again, sorry for dumb questions, I'm just tryin to connect all of these pieces. Which, you know, are A LOT ahah
and thanks for the incredible support you're providing me
btw I probably need to keep diving
It is definitely a LOT haha no worries