Find answers from the community

Updated 2 years ago

Anyone using llamaindex as a means to do

At a glance

The community member is interested in using llamaindex for indexing and search, but is unsure about the changes required when using an OpenAssistant backend instead of OpenAI. The comments suggest referring to the llamaindex documentation on customizing the language model, specifically the section on using a Hugging Face language model. Some example code snippets are also provided, but there is no explicitly marked answer to the community member's question.

Useful resources
Anyone using llamaindex as a means to do indexing and search together with an OpenAssistant backend? Although i've tinkered with llamaindex and VectorIndex to build doc-Q&A with OpenAI, I'm not entirely clear on what parts of that to change when using an Open Assistant model in the backend.. (aosst-sft-4-pythia-12b-epoch-3.5).. Any online examples or snippets floating around ?
L
3 comments
You might be interested in this section of the docs.

I think for open assistant, you'll just want a query wrapper prompt that looks like
<|prompter|>{query_str}<|endoftext|><|assistant|>
Add a reply
Sign up and join the conversation on Discord