HuggingFaceEndpoint
LLM, but am running into issue with the token count. I tried to set the paramater chunk_size
in the service context, but it appears that this is not reducing the amount of tokens.I was wondering if there was anyway to reduce the token/ manage the amount of tokens used in a call? Any help would be appreciated.