OpenAI Rolls Out New Flagship Model, GPT-4o
Overhauling its audio capabilities and improving non-English language usage.
OpenAI is continuously evolving its AI technology to launch new models. In 2024, there are countless chatbots on the market and most companies – like Google, Anthropic and now Microsoft – are developing multiple models built for different usages rather than just evolving a single flagship model
OpenAI has followed suit with GPT-4o, a step up from last year’s GPT-4 model, which debuted in March 2023 as a paid chatbot offered via the ChatGPT Plus membership.
With GPT-4o, “o” stands for “omni,” a nod to what OpenAI calls a nod to a “much more natural human-computer interaction.” The new model is able to take any combination of text, imagery and audio, analyze it and then respond with any combination of the three aforementioned mediums.
AI chatbots have come a long way since just a few years ago. OpenAI says that GPT-4o can also respond to audio in as little as 232 milliseconds (one-fifth of a second).
In addition to better vision and audio analysis, the bot was also revamped from last year’s version to better respond in non-English languages. It seems that the bot still has more work to be done in certain areas.
Despite the abilities OpenAI is advertising, the company is only releasing a version of GPT-4o that allows for text and image inputs and text outputs. Those interested in using it to sift through audio will have to wait.
GPT-4o will be available in the free tier, though Plus members will be allotted five times the message limit.