Azure OpenAI monitoring data reference
See Monitor Azure OpenAI for details on the data you can collect for Azure OpenAI in Microsoft Foundry Models and how to use it.Supported metrics for Microsoft.CognitiveServices/accounts
Here are the most important metrics we think you should monitor for Azure OpenAI. Later in this article is a longer list of all available metrics for this namespace which contains more details on metrics in this shorter list. Please see below list for most up to date information. We’re working on refreshing the tables in the following sections.- Azure OpenAI Requests
- Active Tokens
- Generated Completion Tokens
- Processed FineTuned Training Hours
- Processed Inference Tokens
- Processed Prompt Tokens
- Provisioned-managed Utilization V2
- Prompt Token Cache Match Rate
- Time to Response
- Time Between Tokens
- Time to Last Byte
- Normalized Time to First Byte
- Tokens per Second
- Blocked Volume
- Harmful Volume Detected
- Potential Abusive User Count
- Safety System Event
- Total Volume Sent for Safety Check
The Provisioned-managed Utilization metric is now deprecated and is no longer recommended. This metric has been replaced by the Provisioned-managed Utilization V2 metric.
Tokens per Second, Time to Response, Time Between Tokens are currently not available for Standard deployments.
- ApiName
- FeatureName
- ModelDeploymentName
- ModelName
- ModelVersion
- OperationName
- Region
- StatusCode
- StreamType
- UsageChannel
Supported resource logs for Microsoft.CognitiveServices/accounts
Azure OpenAI microsoft.cognitiveservices/accounts
Related content
- See Monitor Azure OpenAI for a description of monitoring Azure OpenAI.
- See Monitor Azure resources with Azure Monitor for details on monitoring Azure resources.