Find answers from the community

Updated last year

Anyone else having issues loading the

Anyone else having issues loading the 70B llama2 model on LlamaCPP? I was successful with the 7B and 13B models but I’m getting a vague error for 70B. (See attached image)

My cluster is CPU only but has up to 96 workers and 768GB ram.
Attachments
image0.jpg
image1.jpg
L
c
11 comments
Is there anything useful in that assertion error? Looks like it's cutoff
oh wait nvm haha
Is the model path correct?
Seems like somehow the model is not being loaded 🤔
I removed the real path in the photo but I used the right path when I ran into that error
It’s the same dir as the other small models
Let me debug some more before you spend any time on it, you’ve been super human in terms of answering all out questions lol
haha sounds good! Let me know if you find any clues!

I actually haven't tried to run the 70B model with llama.cpp yet, not enough RAM on my end 😉
Found the solution! The bloke himself said to set "n_gqa": 8 for the 70B model. And it worked!
Attachment
image0.jpg
That's so random hahaha nice!
Add a reply
Sign up and join the conversation on Discord