The community member is asking how to change the maximum number of tokens generated by a local large language model (LLM). In the comments, another community member suggests using the additional_kwargs parameter of the Ollama model and setting num_predict to 256 to change the maximum number of generated tokens. There is no explicitly marked answer in the comments.