Log in
Log into community
Find answers from the community
View all posts
Related posts
Did this answer your question?
π
π
π
Powered by
Hall
Inactive
Updated last year
0
Follow
Prompting
Prompting
Inactive
0
Follow
w
woojim
last year
Β·
Yeah from experience using 4bit llama2 70b it's hallucinating pretty badly. Gotta figure out how to prompt better or to finetune it
L
w
7 comments
Share
Open in Discord
L
Logan M
last year
Are you using the INST and EOS/BOS tokens when prompting?
L
Logan M
last year
The format for llama2-chat is pretty strict I've found
w
woojim
last year
Did you mean this?
w
woojim
last year
<s>[INST] <<SYS>>
{{ system_prompt }}
<</SYS>>
{{ user_message }} [/INST]
w
woojim
last year
I was planning to look into incorporating this into llama index calls to a llama 2 api using this format
w
woojim
last year
Not sure if llama index already has this supported format in the code
L
Logan M
last year
We have some utils functions that help with this. If you are implementing a custom llm class, then they will be useful
https://github.com/jerryjliu/llama_index/blob/main/llama_index/llms/llama_utils.py
Add a reply
Sign up and join the conversation on Discord
Join on Discord