Finally, for one example query, I counted ~5k tokens for the above prompt chain ($0.01 with turbo). However, the final cost seemed to be in the $0.30-$0.40 range. Any idea what I'm missing from the final token count?
I have found similar case where the logs says like 1000 token but then you look at the open ai panel and it’s usually 2X sometimes even 5X higher than the reported token in llama. Not sure if it’s on my end but it seems like @aleks_wordcab is experiencing something similar.