The post asks whether an API call is needed to get the token count or if it's an estimation algorithm that runs locally. A community member responds that the token counts reported in the terminal are the actual token counts of data sent to the LLM/Embedding models. They also mention that you can estimate the token usage before running something using a resource from the GPT-Index documentation.