Physical Address
304 North Cardinal St.
Dorchester Center, MA 02124
Physical Address
304 North Cardinal St.
Dorchester Center, MA 02124
OpenAI launches the new GPT-4o mini model on Thursday. The latest mini model is designed to balance the power of GPT-4o with a more cost-efficient approach. This was announced through a blog post.
OpenAI revealed that the 4o mini is much more powerful than other models such as Gemini Flash, Claude Haiku, and GPT-3.5 Turbo.
It surpasses all of them in benchmarks and produces results that are nearly as good as GPT-4o.
The small model is aimed at textual intelligence and multimodal reasoning, the company revealed in a press release.
GPT-4o mini is better than other small models at reasoning tasks, excels in mathematical reasoning and coding, and even performs impressively for multimodal reasoning.
GPT-4o mini enables a broad range of tasks with its low cost and latency, such as applications that chain or parallelize multiple model calls (e.g., calling multiple APIs).
It also passes a large volume of context to other models (e.g., full code base or conversation history).
The model can interact with customers through fast, real-time text responses (e.g., customer support chatbots).
GPT-4o mini also supports text and vision in the API, with support for text, image, video, and audio inputs and outputs coming in the future.
The OpenAI new GPT-4o mini model with a context window of 128K tokens, supports up to 16K output tokens per request and knows up to October 2023. Thanks to the improved tokenizer shared with GPT-4o, handling non-English text is now even more cost-effective.
Also Read: OpenAI’s Upcoming Model May Excel at Reasoning
The company highlighted that GPT-4o Mini excels in mathematical reasoning and coding. However, it has also demonstrated skills in tasks that require reasoning.
OpenAI said in the blog post, “We envision a future where models become seamlessly integrated into every app and on every website.”
“GPT-4o mini is paving the way for developers to build and scale powerful AI applications more efficiently and affordably.”
GPT-4o mini is available as a text and vision model in the Assistants API, Chat Completions API, and Batch API. The fine-tuned version will be rolled out in the coming days.
In ChatGPT, Free, Plus, and Team users will be able to access GPT-4o mini starting today, in place of GPT-3.5.
Enterprise users will also have access soon. This is in line with the company’s mission to make the benefits of AI accessible to all.