Log in
Log into community
Find answers from the community
View all posts
Related posts
Did this answer your question?
π
π
π
Powered by
Hall
Inactive
Updated last year
0
Follow
Quantized LLama2
Quantized LLama2
Inactive
0
Follow
C
Cool
last year
Β·
Hello Experts,
How to use llama_index with quantized llama2 models?
E
L
5 comments
Share
Open in Discord
E
Emanuel Ferreira
last year
https://gpt-index.readthedocs.io/en/latest/examples/llm/llama_2_llama_cpp.html
E
Emanuel Ferreira
last year
Almost the same thing, but then you would use a Hugging Face quantized model
https://huggingface.co/TheBloke/Llama-2-7B-GGML
E
Emanuel Ferreira
last year
maybe @Logan M can validate this
L
Logan M
last year
Yea, you'll want to use llama cpp for ggml or gguf files
https://gpt-index.readthedocs.io/en/stable/examples/llm/llama_2_llama_cpp.html
Huggingface also supports normal quantization using bitsandbytes or gptq
E
Emanuel Ferreira
last year
Quantized LLama2
Add a reply
Sign up and join the conversation on Discord
Join on Discord