Find answers from the community

s
F
Y
a
P
Updated last year

Or i can use the node parser through the

Or i can use the node parser through the service context?
L
A
12 comments
What are you trying to do exactly? A little confused

Yes though, you can use the node_parser on the service context to break documents into nodes
below what I got:
This model's maximum context length is 8192 tokens. However, your messages resulted in 14620 tokens. Please reduce the length of the messages.
And that's why when I ingest, there isn't any textsplitter. Like I said, the textsplitter is in node_parser. But I got again the same error even if I added the nodeparser in the service_context
i supposed that the text splitting happened during "query_time"...now im just trying the reranker but doesnt work for this reason
How did you create the index? Did you use document objects and from_documents() and insert() functions?

It could be a language thing causing the documents not to split well into nodes. You could change to use the recursive character text splitter instead if so
Ive always used tokentextsplitting() and it worked with pre previous version. Btw I'm using Node() and then insert(). Should I change to from_documents?
I would prefer the Node class since I'm more flexible with metadata...
You can keep using the node class yes, it will just be more work to ensure nodes aren't too long haha

How are you creating nodes? Are you splitting text before creating them?
No! And I suspect I should do that. When I used to use Document(), insert() with GPTPineconeIndex the splitting was in automatic. Now I miss this step ahahah
Yes exactly! πŸ™‚
@Logan M sorry, now do you confirm is not anymore in automatic?
Yea, if you create nodes directly, the splitting is not automatic. You'll want to call text_splitter.split_text_with_overlaps(text) before creating the nodes
Add a reply
Sign up and join the conversation on Discord