llm = OpenAI(..., callback_manager=CallbackManager([token_counter]))
>>> from llama_index.callbacks import TokenCountingHandler, CallbackManager >>> tk = TokenCountingHandler() >>> cb = CallbackManager([tk]) >>> from llama_index.llms import OpenAI >>> llm = OpenAI(callback_manager=cb) >>> res = llm.complete("Hello!") >>> tk.total_llm_token_count 11 >>> res = llm.stream_complete("Hello!") >>> for chunk in res: ... continue ... >>> tk.total_llm_token_count 22 >>>