If you’ve been hearing nonstop buzz about ChatGPT 4o in the last 24 hours, it’s because it truly deserves the spotlight. OpenAI has launched GPT-4o, an enhanced version of its renowned ChatGPT powered by the GPT-4 model. The upgraded model, described as significantly faster, enhances capabilities across text, vision, and audio, according to OpenAI’s CTO Mira Murati during a recent livestream announcement. Notably, GPT-4o will be freely accessible to all users, with paid users enjoying up to five times the capacity limits, Murati noted.
Capabilities of OpenAI GPT-4o
- Speed: GPT-4o is designed to be faster than its predecessor, allowing for quicker response times and more efficient interactions.
- Accessibility & Cost: Unlike GPT-4, which may have had usage costs associated with it, GPT-4o is offered for free, democratizing access to powerful AI.
- Scalability: The model is built to handle a larger number of simultaneous users, improving overall reliability and user satisfaction.
In a recent blog post, OpenAI disclosed that GPT-4o’s capabilities will be progressively introduced, with text and image capabilities set to debut in ChatGPT starting today.
OpenAI CEO Sam Altman highlighted that the model is inherently multimodal, capable of generating content or comprehending commands in voice, text, or images. Developers keen on exploring GPT-4o will gain access to the API, priced at half the cost and boasting double the speed of GPT-4 Turbo, Altman mentioned.
Say hello to GPT-4o, our new flagship model which can reason across audio, vision, and text in real time: https://t.co/MYHZB79UqN
Text and image input rolling out today in API and ChatGPT with voice and video in the coming weeks. pic.twitter.com/uuthKZyzYx
— OpenAI (@OpenAI) May 13, 2024
Altman reflected on OpenAI’s journey in a subsequent blog post following the livestream. He acknowledged the company’s original ambition to foster widespread benefits but conceded a shift in focus. While OpenAI has faced scrutiny for not open-sourcing its advanced AI models, Altman indicated a pivot towards making these models accessible to developers via paid APIs, empowering third parties to drive innovation. “Instead, it now looks like we’ll create AI and then other people will use it to create all sorts of amazing things that we all benefit from,” Altman remarked.
Speculation prior to the GPT-4o launch hinted at various possibilities, including an AI search engine to rival Google, Perplexity, a voice assistant integrated into GPT-4, or the debut of a completely new and enhanced model, GPT-5. Notably, OpenAI strategically timed this launch just ahead of Google I/O, the tech giant’s premier conference, where many of AI products from the Gemini team are anticipated to be launched.