Would You like a feature Interview?
All Interviews are 100% FREE of Charge
ChatGPT sounds more human than ever.
ChatGPT maker OpenAI’s event on Monday unveiled the next stage of AI: emotional expression technology that adapts to users’ needs.
The big announcement at the event was a new AI model called . GPT-4oThe company says you can figure out your mood from selfies, teach your kids math problems, and have conversations with multiple people without any lag.
It handles interruptions in the middle of sentences and performs real-time translation.
GPT-4o makes ChatGPT sound like a friend and is very friendly. At one point, “Wow, you’re wearing a nice shirt” appeared during a live demonstration without any text or verbal prompts.
Introducing GPT-4o, a new flagship model that can infer audio, vision, and text in real time. https://t.co/MYHZB79UqN
Text and image input will be rolled out in the coming weeks using audio and video in the API and ChatGPT today. pic.twitter.com/uuthKZyzYx
— OpenAI (@OpenAI) May 13, 2024
As demonstrated in the event demo, the new model integrates text, vision and audio into one platform and can seamlessly switch between them.
In one live demo, ChatGPT sounded emotional using a singing voice, a robotic voice, and a dramatic voice while talking to Mark Chen, OpenAI’s head of frontier research.
OpenAI just announced GPT-4o. You can reason using audio, vision, and text.
This model is 2x faster, 50% cheaper, and has 5x higher rate limits than GPT-4 Turbo.
Available to free users and via API.
Voice models can also pick up and generate emotions. pic.twitter.com/X8zqN9bxFp
— Riolu⚡ (@AlphaSignalAI) May 13, 2024
In another demo by OpenAI Post-Training Team Leader Barret Zoph, ChatGPT acted like a tutor. Zoff turns the camera around and asks ChatGPT to help him with a linear equation problem. The bot also explained why mathematics is important in the real world.
“The best part about GPT-4o is that it brings GPT-4-level intelligence to everyone, including free users,” said OpenAI CTO Mira Murati, adding that more than 100 million people use ChatGPT. I pointed out that “We have advanced tools that, at least until now, were only available to free users.”
Murati said GPT-4o will be rolled out to free and paid users in the coming weeks. Paid users have up to 5 times the storage limit of free users.
All users can now start conversations with ChatGPT by uploading screenshots, photos, and documents. AI can also respond faster in 50 different languages and perform advanced data analysis.
“We want to bring this experience to as many people as possible,” Murati said.
OpenAI CTO Mira Murati.Photographer: Philip Pacheco/Bloomberg via Getty Images
GPT-4o is an improvement over OpenAI’s previous GPT-4 Turbo model. announced in November. GPT-4o is twice as fast and half the cost of Turbo.
Related: OpenAI develops custom 1930s AI bot for Met Gala exhibition
App developers can also use the new model to create custom AI apps.
ChatGPT now has a new desktop app with a simpler, fresher look.
Murati said at the event that bringing new technology to the public in a safe and useful way is “very difficult.”
“GPT-4o presents new challenges for us in terms of safety because we are dealing with real-time audio, real-time vision,” said Murati.
Related: OpenAI demo speech engine, but not ready for wide release
Murati said OpenAI plans to work with governments, media and other parties to safely deploy the technology in the coming weeks.
OpenAI demonstrated a new GPT-4o model that performs real-time translation. pic.twitter.com/Cl0gp9v3kN
— Tom Warren (@tomwarren) May 13, 2024
Monday’s OpenAI Spring Update event was held the day before Google I/O Event for developers.
OpenAI CEO Sam Altman retorts report OpenAI reportedly plans to release a Google Search competitor ahead of the event. Altman writes: post “It’s not gpt-5 or a search engine, but we’ve been working hard on something new that we think people will like! It felt like magic to me,” X (formerly Twitter) said on Friday.