GA Azure OpenAI Emit Token Metric policy in Azure API Management
Share
Services
We're thrilled to announce the General Availability of the Azure OpenAI Emit Token Metric policy in Azure API Management! This new feature enables customers to effortlessly configure token usage metrics to be sent directly to Azure Application Insights, providing insights into the utilization of Azure OpenAI models across various applications and teams.
With this feature, customers can easily set up a policy to transmit token usage metrics to Application Insights with a straightforward configuration. The policy captures prompt, completions, and total token usage metrics, offering a comprehensive view of model utilization.
Customers have the flexibility to specify the Application Insights namespace to which metrics should be sent, ensuring seamless integration with their existing monitoring infrastructure. Moreover, they can configure or select from pre-defined dimensions to segment token usage metrics, enabling granular analysis by Subscription ID, IP Address, User ID, and more.
This new policy streamlines the process and supports both streaming and non-streaming scenarios for Azure OpenAI. It empowers customers to gain deeper insights into token consumption effortlessly, facilitating better decision-making and resource allocation.
Click [here](https://aka.ms/apim/openai/token-metric-policy) to learn more.
* API Management
* Features
* Microsoft Build
* [ API Management](https://azure.microsoft.com/en-gb/products/api-management/)
What else is happening at Microsoft Azure?
Read update
Services
Share
Generally Available: Storage account default maximum request rate limit increase to 40,000 requests per second
December 12th, 2024
Services
Share
Read update
Services
Share
Generally Available: Regional Disaster Recovery by Azure Backup for AKS
November 22nd, 2024
Services
Share
Generally Available: Enhancements on Azure Container Storage for performance, scalability, and operational insights
November 19th, 2024
Services
Share