ChatGPT sounds more human than ever.
An event Monday by ChatGPT creator OpenAI unveiled the next phase of AI: Emotionally expressive technology that adapts to users' needs.
The big announcement at the event was a new AI model called GPT-4owhich the company says can understand how you feel from a selfie, teach kids through math problems and talk to multiple people without lag.
It can even handle interruption in the middle of a sentence and perform real-time translations.
GPT-4o makes ChatGPT sound like a friend — a super friendly friend. At one point, it said, “Wow, that's a nice shirt you're wearing,” in direct demonstrations, without texting or verbal prompting.
Say hello to GPT-4o, our new flagship model that can reason in real-time audio, vision and text: https://t.co/MYHZB79UqN
Text and image input will be rolling out to the API today and ChatGPT with voice and video in the coming weeks. pic.twitter.com/uuthKZyzYx
— OpenAI (@OpenAI) May 13, 2024
The new model brings together text, vision and audio on one platform and can be seamlessly transitioned, as shown by the demos at the event.
In a live demo, ChatGPT sounded emotive using a singing voice, a robotic voice, and a dramatic voice when talking to Mark Chen, OpenAI's head of frontier research.
OpenAI just announced “GPT-4o”. Can reason with voice, vision and text.
The model is 2x faster, 50% cheaper and has 5x higher rate limit than GPT-4 Turbo.
It will be available to users for free and via API.
The voice model can even understand emotions and generate… pic.twitter.com/X8zqN9bxFp
— Lior✓ (@AlphaSignalAI) May 13, 2024
In another demo, this one by OpenAI post-training team leader Barret Zoph, ChatGPT acted as a tutor. Zoph turned on his camera and asked ChatGPT to help him with a linear equation problem. The bot even explained why math mattered in the real world.
“The best thing about GPT-4o is that it brings GPT-4 level intelligence to everyone, including our free users,” said OpenAI CTO Mira Murati, noting that more than 100 million people use ChatGPT. “We have advanced tools that have only been available to free users, at least until now.”
Murati said GPT-4o will be rolled out to free and paid users in the coming weeks. Paying users will have up to five times the capacity limit of free ones.
All users can now upload screenshots, photos and documents to start chats with ChatGPT. AI will also respond faster, in 50 different languages, and can perform advanced data analysis.
“We want to bring this experience as much as possible”, said Murati.
CTO of OpenAI Mira Murati. Photo: Philip Pacheco/Bloomberg via Getty Images
GPT-4o is an improvement on OpenAI's previous GPT-4 Turbo model, which has announced in November. GPT-4o is twice as fast and half as expensive as Turbo.
Connected: OpenAI Develops Custom 1930s AI Bot for Met Gala Exhibition
App developers can also use the new model to make custom AI apps.
ChatGPT is also getting a new desktop app and a simpler, refreshed look.
Murat stated at the event that it was “quite challenging” to bring the new technology to the public in a safe and useful way.
“GPT-4o presents new challenges for us when it comes to security because we're dealing with real-time audio, real-time vision,” Murati said.
Connected: OpenAI Demos Voice Engine, but not ready for wide release
According to Murat, OpenAI is working with governments, media and other entities to deploy the technology securely in the coming weeks.
OpenAI has just demonstrated its new GPT-4o model doing real-time translations? pic.twitter.com/Cl0gp9v3kN
– Tom Warren (@tomwarren) May 13, 2024
The OpenAI Spring Update event on Monday happened a day ago Google I/O event for developers.
OpenAI CEO Sam Altman denied it REPORTS that OpenAI would launch a Google search competitor ahead of the event. Altman wrote in one POST on X, formerly Twitter, on Friday: “no gpt-5, not a search engine, but we've been working hard on some new stuff we think people will love! Sounds like magic to me.”