Requirements
Organization Admin permissions are required to access the Token Tracking dashboard.What are tokens?
Tokens are the unit of measurement for AI model usage. Every interaction with an AI feature in Omni — such as asking a question, generating a summary, asking AI to make data model changes, or using the MCP Server — consumes tokens. Token usage varies depending on several factors:- Task complexity — More complex questions, multi-step analysis, or larger datasets use more tokens than simple lookups. Each message in a session carries prior context, so long running sessions can compound.
- Data model and context size — Omni sends context from the data model to the LLM to improve answer accuracy (field descriptions, ai_context, synonyms, etc). Larger models use more tokens.
- LLM model — More capable models (e.g., Sonnet-class) consume more tokens per request than lighter models (e.g., Haiku-class). See AI model settings to configure tiers.
Using the Token Tracking dashboard
To access the Token Tracking dashboard, click Analytics in the left navigation of the Omni app and select the Token Tracking dashboard. The dashboard displays token consumption over time and can break down usage in various ways, such as by feature and user.Optimizing token usage
- Adjust the LLM model tier. See AI model settings.
- Trim unused fields, descriptions, or
ai_contextfrom the data model. Reducing context may lower per-call token counts, but can increase the number of turns needed to get a good answer. The goal is optimizing token efficiency overall.
FAQ
Why does a single prompt generate multiple rows in the tracking data?
Why does a single prompt generate multiple rows in the tracking data?
Omni’s AI is an agentic system. For a given prompt or request, it may fire off multiple tool calls to get the answer or complete the task. For example, asking “show me revenue by region” might involve the AI searching the data model for the right fields, building the query, creating a visualization, and summarizing results — each as a separate step. Each of these steps is logged individually in the tracking data, which is why a single user prompt can appear as several rows.
What are the blank prompts with token usage?
What are the blank prompts with token usage?
These are “tool calls” — intermediate steps where the AI determines which action to take next. They don’t have a user-facing prompt because they are part of the AI’s internal reasoning as it works through a request. See the question above for more context on why these occur.
Why does the same question sometimes use different amounts of tokens?
Why does the same question sometimes use different amounts of tokens?
Token usage can vary between identical prompts due to differences in conversation context (earlier messages in the session) or the AI choosing a different path to arrive at the answer.
Do MCP calls consume tokens?
Do MCP calls consume tokens?
Yes, MCP calls do consume tokens. The granular tracking data for an individual MCP call may show 0 tokens, but the token usage is logged under the associated query tool call.
Next steps
- AI settings — Configure AI features and model providers
- AI data privacy — Learn how Omni’s AI features process data
- Usage analytics — Explore other pre-built analytics dashboards
- Deliveries and alerts - Set up alerts to be notified about token usage