Find answers from the community

Updated last year

I m on a MacBook Pro M2 Max and when I

I'm on a MacBook Pro M2 Max, and when I run index.as_query_engine().query('question') I get this error:

Plain Text
RuntimeError: MPS does not support cumsum op with int64 input
L
p
6 comments
I get that too πŸ˜… (I have the same macbook)

Never bothered to fix it -- i just use llama.cpp instead, compiled with apple metal support
I think it has something to do with installing torch with apple silicon support? Not sure though
Got it.

I'm a rookie, so very rookie question coming up:

any tips/guides on how to integrate llama.cpp with llama index? is it just getting the model locally via llama.cpp and referencing (how?) in llama index?

thanks again for all the effort and work you put into this!
Right! We have a guide here on llama.cpp

https://gpt-index.readthedocs.io/en/stable/examples/llm/llama_2_llama_cpp.html

This will download and run llama2-13B-chat locally πŸ™‚
To save you time, here's the install command for macs
CMAKE_ARGS="-DLLAMA_METAL=on" FORCE_CMAKE=1 pip install llama-cpp-python
Add a reply
Sign up and join the conversation on Discord