Does azure-openai-emit-token-metric policy at API Management Service support cached tokens?
The Azure API Management Service recently updated the azure-openai-emit-token-metric
to support the GPT-4o model.
Does this policy support the recording of Cached Tokens? According to the official documentation on MS Learn, it appears that this feature is not supported.
https://learn.microsoft.com/en-us/azure/api-management/azure-openai-emit-token-metric-policy
Token count metrics include: Total Tokens, Prompt Tokens, and Completion Tokens.
However, I’m unsure if this is due to the documentation not being updated or if the feature is indeed not supported at all.
Feel free to ask if you need any further assistance!