On Thursday, OpenAI launched its new AI model, the “GPT-4o mini”. This AI model is the startup’s latest attempt to expand the use of its chatbot.
According to OpenAI, this is the most capable and cost-efficient small model available today, with plans to integrate image, video, and audio capabilities.
GPT-4o Mini is an offspring of GPT-4o, the fastest and most powerful model launched by OpenAI in May during a livestream. The startup explains that the “o” in GPT-4o stands for omni, highlighting the improved audio, video, and text capabilities of the GPT-4o, which can handle 50 different languages at high speeds.
Founded in 2015, the AI startup OpenAI has been on top of the generative AI market since its inception as it continues to spend colossal sums of money on infrastructures and processors to develop and train its AI models.
The world is multimodal. If you think about the way we as humans process the world and engage with the world, we see things, we hear things, we say things – the world is much bigger than text. So to us, it always felt incomplete for text and code to be the single modalities, the single interfaces we could have to how powerful these models are and what they can do.
Brad Lightcap, OpenAI Chief Operating Officer
GPT-4o mini is now available to all users of ChatGPT except ChatGPT Enterprise users until next week.