Elevated AI Total Tokens

firetail:ai-total-tokens-elevated

Type:

Detection

Rule Severity:

Info

The total number of input and output tokens processed has increased significantly.

This could result in higher costs, increased latency, and potential throttling of requests.

Potential Causes:

  • Large inputs or inefficient prompts leading to excessive token consumption.
  • Model generating overly detailed responses.
  • Misuse of AI service leading to cost spikes.

Remediation

Monitor usage patterns, make adjustments to model configurations accordingly to ensure optimal performance.

Example Attack Scenario

How to Identify with Example Scenario

How to Resolve with Example Scenario

How to Identify with Example Scenario

Find the text in bold to identify issues such as these in API specifications

How to Resolve with Example Scenario

Modify the text in bold to resolve issues such as these in API specifications
References:

More findings

All Findings