Azure OpenAI Metrics Issue: Unable to Track Assistant Inference Tokens

Andrea ESPOSITO 5 Reputation points
2025-01-16T13:20:05.1666667+00:00

Hello,

I am encountering an issue with Azure OpenAI Metrics: I can no longer track the number of inference tokens generated by a gpt model when I use OpenAI assistants under the metric AZURE OPENAI - USAGE > Processed Inference Tokens.

Previously, these tokens were correctly reported on the dashboard, but now it seems that inference token data is either missing or not updated.

To give more context, if I call a gpt model through Azure AI Foundry Chat Playground I can track the Processed Inference Token as they are reported. If I call the same gpt model through Azure AI Foundry Assistant Playground the tokens generated are not reported.

So

  1. Has there been a recent change in how Azure tracks metrics for Assistant models?
  2. Is there a configuration or workaround to continue monitoring inference tokens?
  3. Could this be a known bug, or does it require configuration changes on my end?

Thank you in advance for your support!

Azure OpenAI Service
Azure OpenAI Service
An Azure service that provides access to OpenAI’s GPT-3 models with enterprise capabilities.
3,784 questions
0 comments No comments
{count} vote

1 answer

Sort by: Most helpful
  1. SriLakshmi C 3,245 Reputation points Microsoft External Staff
    2025-01-16T21:42:32.6+00:00

    Hello Andrea ESPOSITO,

    Greetings and Welcome to Microsoft Q&A! Thanks for posting the question.

    I understand that you are encountering an issue with Azure OpenAI Metrics. I attempted to reproduce the issue but was unable to encounter any problems. The metrics appear to be displayed correctly. Here is the screen shot,

    This is for Azure AI Foundry Chat Playground,User's image

    This is for Azure AI Foundry Assistant Playground,

    User's image

    Has there been a recent change in how Azure tracks metrics for Assistant models?

    As of now, there has been no official announcement regarding changes in how Azure tracks metrics for Assistant models. However, discrepancies in metric reporting can occur between different tools, such as the Azure AI Foundry Chat Playground and the Assistant Playground. These differences may result from variations in how usage data is logged and reported by the platforms.

    Is there a configuration or workaround to continue monitoring inference tokens?

    Here are a few workarounds:

    • Ensure Diagnostic Settings Configuration is enabled for your Azure OpenAI resource and includes relevant metrics like "Processed Inference Tokens" in the logs.
    • Leverage Log Analytics Integration by connecting your Azure OpenAI resource to Log Analytics. This allows you to capture detailed logs and use custom queries to track token usage.
    • Implement API-Level Monitoring by capturing request and response payloads when using the API and calculating token usage locally using OpenAI’s token estimation methods.

    Could this be a known bug, or does it require configuration changes on my end?

    The issue might be a configuration error. Ensure Diagnostic Settings are set up correctly to track "Processed Inference Tokens" and routed to Log Analytics or other destinations.

    Kindly refer this Monitor Azure OpenAI and Azure OpenAI monitoring data reference.

    I hope you understand. And, if you have any further query do let us know.


    If this answers your query, do click Accept Answer and Yes for was this answer helpful.

    Thank you!


Your answer

Answers can be marked as Accepted Answers by the question author, which helps users to know the answer solved the author's problem.