OpenAI Unveils New Cost-Efficient GPT-4o Mini
OpenAI is committed to making intelligence as broadly accessible as possible. Today (18th July), they announced the GPT-4o Mini, their most cost-efficient small model. This model is expected to significantly expand the range of applications built with AI by making intelligence much more affordable. Priced at 15 cents per million input tokens and 60 cents per million output tokens, GPT-4o mini is more than 60% cheaper than GPT-3.5 Turbo and an order of magnitude more affordable than previous frontier models.
Versatile and Powerful: Capabilities of GPT-4o Mini
GPT-4o mini enables a broad range of tasks with low cost and latency. This includes applications that chain or parallelize multiple model calls, pass a large volume of context to the model, or interact with customers through fast, real-time text responses. Currently, it supports text and vision in the API, with plans to include text, image, video, and audio inputs and outputs. It has a context window of 128K tokens and supports up to 16K output tokens per request, with knowledge up to October 2023.
GPT-4o mini surpasses GPT-3.5 Turbo and other small models across textual intelligence and multimodal reasoning on academic benchmarks. It scores 82% on MMLU, 87% on MGSM for math reasoning, and 87.2% on HumanEval for coding performance. Additionally, it demonstrates strong performance in function calling, enabling developers to build applications that fetch data or take actions with external systems.
Collaboration and Real-world Applications
OpenAI has partnered with companies like Ramp and Superhuman to understand the use cases and limitations of GPT-4o mini. These companies found GPT-4o mini to perform significantly better than GPT-3.5 Turbo for tasks such as extracting structured data from receipt files and generating high-quality email responses.
Safety is a priority for OpenAI. GPT-4o mini has the same safety mitigations as GPT-4o, reinforced through pre-training filters and post-training alignment with policies. The model applies a new safety technique called “instruction hierarchy,” making it more resistant to jailbreaks and prompt injections.
Availability and Pricing
GPT-4o mini is available now in the Assistants API, Chat Completions API, and Batch API. Developers pay 15 cents per 1M input tokens and 60 cents per 1M output tokens. It is accessible to Free, Plus, and Team users in ChatGPT, with enterprise access starting next week.
OpenAI envisions a future where AI models become seamlessly integrated into every app and website. GPT-4o mini paves the way for developers to build and scale powerful AI applications more efficiently and affordably, driving down costs while enhancing model capabilities.
Conclusion
OpenAI’s launch of GPT-4o mini is a milestone in making advanced AI accessible and cost-effective. Combining high performance with affordability empowers developers to create impactful solutions. This advancement democratizes AI, promising widespread integration and benefits across various sectors.