Log in
Log into community
Find answers from the community
View all posts
Related posts
Did this answer your question?
😞
😐
😃
Powered by
Hall
Inactive
Updated last year
0
Follow
“Llama generate prefix match hit”
“Llama generate prefix match hit”
Inactive
0
Follow
c
captam_morgan
last year
·
“Llama.generate: prefix-match hit”
L
c
3 comments
Share
Open in Discord
L
Logan M
last year
llama.cpp has some sort of built-in cache, that just means your generation will be faster... I think
Setting verbose=False on the LLM object should hopefully reduce the amount of logging, althogh the llama.cpp library is pretty noisy in general
L
Logan M
last year
It's just a notification though, nothing to worry about 👍
c
captam_morgan
last year
Awesome, thanks!
Add a reply
Sign up and join the conversation on Discord
Join on Discord