The total number of input and output tokens processed has increased significantly.
This could result in higher costs, increased latency, and potential throttling of requests.
Potential Causes:
- Large inputs or inefficient prompts leading to excessive token consumption.
- Model generating overly detailed responses.
- Misuse of AI service leading to cost spikes.