Find answers from the community

Updated 2 years ago

Anyone else having issues loading the

At a glance

A community member is having issues loading the 70B llama2 model on LlamaCPP, while they were successful with the 7B and 13B models. The community members discuss potential causes, such as the model path being incorrect or the model not loading properly. Eventually, one community member finds a solution by setting the "n_gqa" parameter to 8 for the 70B model, which resolves the issue.

Anyone else having issues loading the 70B llama2 model on LlamaCPP? I was successful with the 7B and 13B models but I’m getting a vague error for 70B. (See attached image)

My cluster is CPU only but has up to 96 workers and 768GB ram.
Attachments
image0.jpg
image1.jpg
L
c
11 comments
Is there anything useful in that assertion error? Looks like it's cutoff
oh wait nvm haha
Is the model path correct?
Seems like somehow the model is not being loaded 🤔
I removed the real path in the photo but I used the right path when I ran into that error
It’s the same dir as the other small models
Let me debug some more before you spend any time on it, you’ve been super human in terms of answering all out questions lol
haha sounds good! Let me know if you find any clues!

I actually haven't tried to run the 70B model with llama.cpp yet, not enough RAM on my end 😉
Found the solution! The bloke himself said to set "n_gqa": 8 for the 70B model. And it worked!
Attachment
image0.jpg
That's so random hahaha nice!
Add a reply
Sign up and join the conversation on Discord