Log in
Log into community
Find answers from the community
View all posts
Related posts
Did this answer your question?
π
π
π
Powered by
Hall
Inactive
Updated 4 months ago
0
Follow
whats interesting is that I am able to
whats interesting is that I am able to
Inactive
0
Follow
k
khalid
4 months ago
Β·
whats interesting is that I am able to run the model with ollama command on server directly though
L
k
4 comments
Share
Open in Discord
L
Logan M
4 months ago
Memory will grow until it reaches the max context limit. Its lazily allocated
L
Logan M
4 months ago
Setting a limit on the context window size is the way to limit the memory usage yes
L
Logan M
4 months ago
llm = Ollama(..., context_window=3000)
for example may help limit memory usage, but the lower you put it, the less context you can fit into the llm, which may increase the number of llm calls needed to run a query
k
khalid
4 months ago
I will try this out, thanks for your help.
Add a reply
Sign up and join the conversation on Discord
Join on Discord