May 14, 2024
The most recent version of OpenAI's artificial intelligence model may attempt to identify people's moods and emulate human cadences in spoken answers. This impact reminds me of the 2013 Spike Jonze film "Her," in which the main character (who is human) develops feelings for an artificially intelligent operating system, which leads to some issues. Few people will find the new model alluring, but according to OpenAI, it does operate more quickly than earlier iterations and has real-time reasoning capabilities for text, audio, and video. During a brief live-streamed update, the firm said that GPT-4o, short for "omni," will power OpenAI's well-known ChatGPT chatbot and be available to customers, even those using the free version, in the coming weeks. CEO Sam Altman only tweeted the word "her" on social media platform X; he did not present at the event.
The AI bot conversed in real time during a presentation with Chief Technology Officer Mira Murati and other executives, including emotion—specifically, “more drama”—in response to requests. It also helped with a more complicated software code issue on a computer screen and guided the user through the process of solving a basic math equation without automatically providing the solution.
To demonstrate how technology may facilitate communication between speakers of different languages, it also attempted to infer an individual's emotional state from a selfie video of their face by determining that they were pleased since they were smiling. The results were translated into Italian and English. The less than half-hour update created the impression that OpenAI is catching up to more established competitors, according to Gartner researcher Chirag Dekate.
"We had seen advanced versions of these demos showcased by Google in their Gemini 1.5 pro launch, so many of the demos and capabilities showcased by OpenAI seemed familiar," Dekate added. "We are seeing capability gaps emerge, even though Open AI had a first-mover advantage with ChatGPT and GPT3 last year when compared to their peers, especially Google."
Tuesday and Wednesday are Google's I/O developer conference days. It is anticipated that the company will release improvements to its AI model, Gemini.