Is your text an input prompt or a model's output? This tool helps with both. Paste your text to calculate its token count. Then, see how it performs as an input (predicting the AI's available response room) or analyze it as an output (to check if it was truncated by the model's limit).
Have you noticed your AI's responses aren't quite as long as the total limit it advertises? There's a good reason for that!
When an AI tells you it has a limit of 65,536 tokens, it sounds like you should get that many units of information. But in reality, you'll usually see the actual message you get is closer to 60,000 to 62,000 tokens.
Why the difference? Because the AI uses some of its "token budget" for its own internal work, sort of like background thinking. These hidden tokens are used for things like:
So, while the AI has a big maximum limit, a part of it is always used for its own operations. What you see is the usable output after all that internal work is done.
Important: Once a model hits its maximum output limit, it will stop generating content, which can result in an incomplete response. To avoid this, it is critical to ensure your requested output fits well within the limit. This tool's 'Output Truncation Analysis' will change to Approaching Limit at 70% and High Risk when your content exceeds 90% of the selected model's limit, signaling a high probability of truncation.