GPT-4o mini can now be deployed in the East US Azure region, with more regions coming soon. Having been available in preview for the last week via the Early Access Playground, partners and customers can now access GPT-4o mini via the Azure OpenAI API.
This new, smaller version of OpenAI's GPT-4o model offers speed, capability, vision, 128k context window, all at a fraction of the cost of the standard GPT-4o model - and even cheaper than GPT-3.5 Turbo.
GPT-4o mini will initially be available in Standard Global, and Standard Regional deployment types, with Batch, and Provisioned coming soon. Standard Global pricing is a staggering $0.15 per 1M tokens for input, and $0.60 per 1M tokens for output.
Based on my own experience using GPT-4o mini in the Azure OpenAI Playground, and also how strongly it's scoring on independent evaluations, I'm convinced that it's going to quickly become the most popular model in the Azure OpenAI Service.
If you want to try GPT-4o mini yourself, try deploying the model in Azure OpenAI Studio, or use one of the many code samples on GitHub. If you've never used the Azure OpenAI Service before, you'll be pleased to know that last month, the signup process was streamlined, removing the need to submit a manual form for access.