Log in
Log into community
Find answers from the community
View all posts
Related posts
Did this answer your question?
π
π
π
Powered by
Hall
Inactive
Updated 3 months ago
0
Follow
Has anyone else noticed HuggingFaceLLMs
Has anyone else noticed HuggingFaceLLMs
Inactive
0
Follow
c
cmagorian
last year
Β·
Has anyone else noticed HuggingFaceLLMs just hanging on their machine once you get to
query
? I have been unable to get a response from them and ultimately have to do a KeyboardInterrupt
L
c
11 comments
Share
Open in Discord
L
Logan M
last year
I've seen this happen if you don't configure the max input size correctly π€ What llm are you using?
c
cmagorian
last year
"StabilityAI/stablelm-tuned-alpha-3b"
c
cmagorian
last year
and im using an 8GB GPU, i know its not much
c
cmagorian
last year
just got the cuda.OutOfMemoryError
c
cmagorian
last year
where do i set the max_input_size? @Logan M
L
Logan M
last year
ah wrong term, I meant context_window
Maybe since you have some limited memory, you can atrtifically lower it to 2048 (both the context_window and tokenizer kwargs)
https://gpt-index.readthedocs.io/en/latest/core_modules/model_modules/llms/usage_custom.html#example-using-a-huggingface-llm
L
Logan M
last year
You probably already saw the demo from that page haha
c
cmagorian
last year
ah gotcha, i just tried 1024, but still had CUDA OOM, will try CPU albiet slow to see if that works
c
cmagorian
last year
Is there a way to reduce batch_size?
L
Logan M
last year
Batch size should already be 1 π
ngl 8GB is tough to work with. And tbh, open source LLMs in general are still npt great, especially the smaller ones
c
cmagorian
last year
yeah i tried a much smaller llm and results werent great
Add a reply
Sign up and join the conversation on Discord
Join on Discord