Find answers from the community

Home
Members
failfast
f
failfast
Offline, last seen 2 months ago
Joined September 25, 2024
i figured out how to increase the max_tokens so that my answers don't get cut off. however, i don't see a way to do this if you load_from_disk your index (without creating the index in the session from scratch)....am i missing something?
3 comments
f
j
Thanks for your inputs! 95% of my documents are powerpoints, so i was planning on chunking slide by slide and generating an embedding per slide. is that the same concept as using sentence transformers?

my main question though is what should the GPT-Index index structure be? because of the vast amount of data, would I need to go in a mult-level tree direction? would this hinder performance?
7 comments
J
f
r
has anyone taken an existing elasticsearch (or any) index and retrofitted llamaindex on top of it? i am trying to do that with an existing elasticsearch index i have, but keep running into

File "C:\Python311\Lib\site-packages\llama_index\vector_stores\elasticsearch.py", line 542, in aquery
f"Could not parse metadata from hit {hit['_source']['metadata']}"
~~~~~~^^^^^^^^^^^^
KeyError: 'metadata'

understandably because the format of the metadata in my Elasticsearch index is not the expected format for the ElasticsearchStore class in LlamaIndex. how do you get around this without having to re-index everything?
3 comments
f
L
hi, i am trying to build a chatbot using openai on azure. i am following these instructions: https://gpt-index.readthedocs.io/en/latest/examples/agent/openai_agent.html. i've replaced this line:

Plain Text
llm: OpenAI = OpenAI(temperature=0, model="gpt-3.5-turbo-0613"),


with this:

Plain Text
llm: AzureOpenAI = AzureOpenAI(engine="gpt-4", model="gpt-4", temperature=0.7, max_tokens=800, top_p=0.95, frequency_penalty=0, presence_penalty=0, stop=None),


but I am getting this error:

Plain Text
openai.error.InvalidRequestError: Unrecognized request argument supplied: functions


in this line:

Plain Text
ai_message = self._llm.chat(chat_history, functions=functions).message


does anything immediately jump out? am i able to simply switch the llm to use AzureOpenAI?
12 comments
L
f
B
does llama index support Vespa?
1 comment
L
curious if anyone here has used gpt-index on top of a truly large corpus of data without training the model on your corpus, and whether the results were production quality? i'm in the process of trying to implement this on my corpus of data (115k docs, ~15k words per doc) using OpenAI, and I'm concerned I will need to change course and use a model I can actually train on my data because relevancy won't be good enough
1 comment
M