Find answers from the community

Updated 2 years ago

Anyone using llamaindex as a means to do

Anyone using llamaindex as a means to do indexing and search together with an OpenAssistant backend? Although i've tinkered with llamaindex and VectorIndex to build doc-Q&A with OpenAI, I'm not entirely clear on what parts of that to change when using an Open Assistant model in the backend.. (aosst-sft-4-pythia-12b-epoch-3.5).. Any online examples or snippets floating around ?
L
3 comments
You might be interested in this section of the docs.

I think for open assistant, you'll just want a query wrapper prompt that looks like
<|prompter|>{query_str}<|endoftext|><|assistant|>
Add a reply
Sign up and join the conversation on Discord