OpenAI on Monday launched a new AI model and desktop version of ChatGPT, along with an updated user interface, the company’s latest effort to expand use of its popular chatbot.

The update brings GPT-4 to everyone, including OpenAI’s free users, technology chief Mira Murati said in a livestreamed event. She added that the new model, GPT-4o, is “much faster,” with improved capabilities in text, video and audio. OpenAI said it eventually plans to allow users to video chat with ChatGPT.

“This is the first time that we are really making a huge step forward when it comes to the ease of use,” Murati said.

OpenAI, backed by Microsoft, has been valued by more than $80 billion by investors. The company, founded in 2015, is under pressure to stay on top of the generative AI market while finding ways to make money as it spends massive sums on processors and infrastructure to build and train its models.

The o in GPT-4o stands for omni. The new model allows ChatGPT to handle 50 different languages with improved speed and quality, and it will also be available via OpenAI’s API so that developers can begin building applications using the new model today, Murati said.

She added that GPT-4o is twice as fast as, and half the cost of, GPT-4 Turbo.

OpenAI team members demonstrated the new model’s audio capabilities, asking for help calming down ahead of a public speech. OpenAI researcher Mark Chen said the model has the capability to “perceive your emotion,” adding that the model can also handle users interrupting it. The team also asked it to analyze a user’s facial expression to comment on the emotions the person may be experiencing.

“Hey there, what’s up? How can I brighten your day today?” ChatGPT’s audio mode said when a user greeted it.

The company plans to test Voice Mode in the coming weeks, with early access for paid subscribers to ChatGPT Plus, according to a blog post. OpenAI also said the new model can respond to users’ audio prompts “in as little as 232 milliseconds, with an average of 320 milliseconds, which is similar to human response time in a conversation.”

CNBC