huggingface_hub.inference._text_generation.ValidationError: Input validation error: `inputs` tokens + `max_new_tokens` must be <= 4096. Given: 3896 `inputs` tokens and 256 `max_new_tokens` make: *** [evals] Error 1
from functools import partial from transformers import AutoTokenizer from llama_index import set_global_tokenizer set_global_tokenizer( partial(AutoTokenizer.from_pretrained("TheBloke/prometheus-13B-v1.0-AWQ").encode, add_special_tokens=False) )