Politics

What to know about the launch of GPT-4o

Share on facebook
Share on twitter
Share on linkedin
Share on pinterest
Share on telegram
Share on email
Share on reddit
Share on whatsapp
Share on telegram



OpenAI on Monday launched its latest artificial intelligence (AI) model, GPT-4o, which promises improvements in its text, vision and audio capabilities.

OpenAI revealed the model during a live demonstration on Monday, with chief technology officer Mira Murati saying it is a “huge advancement in the ease of use” of the system. OpenAI’s newest model was released just one day before Google’s annual developer conference scheduled for Tuesday.

Here’s what you should know about the GPT-4o launch.

Improved voice instruction

Users can now show multiple GPT-4o photos and chat with the model about the uploaded image, according to OpenAI.

This can help students solve math problems step by step. One of the demonstrations shown during the launch on Monday guides users through a simple math problem without providing any answers.

A separate video posted by online learning company Khan Academy demonstrates how the new model can help teach students In real time. The student shared his screen with him working on the problem in real time as the model guided him.

A faster model with improved features

Murati said on Monday that GPT-4o provides “GPT-4-level intelligence” that is faster and improves the system’s capabilities in text, vision and audio.

“This is really changing the paradigm for the future of collaboration, where this interaction becomes much more natural and much, much easier,” she said.

OpenAI said your new model can “respond to audio inputs in just 232 milliseconds, with an average of 320 milliseconds.” He noted that this is about the same time it takes humans to respond in conversation.

The new model was launched on Monday

GPT-4o is available starting Monday to all users of OpenAI’s ChatGPT AI chatbot, including those using the free version.

“GPT-4o’s text and image features are starting to roll out today in ChatGPT. We are making GPT-4o available at the free tier and for Plus users with message limits up to 5x greater”, OpenAI wrote in his update Monday.

The new voice mode will be rolling out in the coming weeks to ChatGPT Plus users, OpenAI CEO Sam Altman wrote on social platform.

The model is ‘natively multimodal’

Altman also posted on X that the model is “natively multimodal,” meaning the model can generate content and understand commands through voice, text, or images.

On a separate blog posthe said the new voice and video mode “is the best computer interface” he has ever used.

“It looks like movie AI; and it’s still a little surprising to me that it’s real. Reaching human-level response times and expressiveness ends up being a huge change,” he wrote in Monday’s post.





This story originally appeared on thehill.com read the full story

Support fearless, independent journalism

We are not owned by a billionaire or shareholders – our readers support us. Donate any amount over $2. BNC Global Media Group is a global news organization that delivers fearless investigative journalism to discerning readers like you! Help us to continue publishing daily.

Support us just once

We accept support of any size, at any time – you name it for $2 or more.

Related

More

1 2 3 6,256

Don't Miss