Hi, I'm trying to use LlamaCPP too, and it works with a small test dataset (just one text file with one sentence !). When I load more data, I always get a
llama_tokenize: too many tokens
error
Here is more context :
`
File ~/.pyenv/versions/3.11.2/lib/python3.11/site-packages/llama_index/indices/base.py:100, in BaseGPTIndex.from_documents(cls, documents, docstore, service_context, **kwargs)
96 docstore.set_document_hash(doc.get_doc_id(), doc.get_doc_hash())
98 nodes = service_context.node_parser.get_nodes_from_documents(documents)
--> 100 return cls(
101 nodes=nodes,
102 docstore=docstore,
103 service_context=service_context,
104 **kwargs,
105 )
File ~/.pyenv/versions/3.11.2/lib/python3.11/site-packages/llama_index/indices/tree/base.py:72, in GPTTreeIndex.__init__(self, nodes, index_struct, service_context, summary_template, insert_prompt, num_children, build_tree, use_async, **kwargs)
70 self.build_tree = build_tree
...
114 if int(n_tokens) < 0:
--> 115 raise RuntimeError(f'Failed to tokenize: text="{text}" n_tokens={n_tokens}')
116 return list(tokens[:n_tokens])
`