HomeArtificial IntelligenceChatGPT now understands real-time video, seven months after OpenAI first demonstrated it

ChatGPT now understands real-time video, seven months after OpenAI first demonstrated it

OpenAI has finally released the real-time video capabilities for ChatGPT that it demonstrated almost seven months ago.

On Thursday, the corporate announced during a livestream that Advanced Voice Mode, its human-like conversation feature for ChatGPT, is getting a vision. The ChatGPT app allows users subscribed to ChatGPT Plus, Team, or Pro to point their phones at objects and have ChatGPT respond in near real time.

Advanced Voice Mode with Vision also can understand what's on a tool's screen through screen sharing. For example, it will probably explain various settings menus or give suggestions for a math problem.

To access advanced voice mode with Vision, tap the voice icon next to the ChatGPT chat bar after which tap the video icon at the underside left to begin the video. To share the screen, tap the three-dot menu and choose Share Screen.

The rollout of Advanced Voice Mode with Vision will begin on Thursday and be accomplished next week, based on OpenAI. However, not all users get access. According to OpenAI, ChatGPT Enterprise and Edu subscribers won’t receive the feature until January and there isn’t a timeline for ChatGPT users within the EU, Switzerland, Iceland, Norway or Liechtenstein.

In one Current demo On CNN's “60 Minutes,” OpenAI President Greg Brockman conducted a vision quiz with Anderson Cooper on his anatomical skills in Advanced Voice Mode. As Cooper drew body parts on a whiteboard, ChatGPT was capable of “understand” what he was drawing.

OpenAI employees display ChatGPT's advanced voice mode with Vision during a livestream. Photo credit:OpenAI

“The location is excellent,” said ChatGPT. “The brain is correct in the pinnacle. As far as form goes, it's a superb start. The brain is more oval.”

However, in the identical demo, Advanced Voice Mode with Vision made a mistake on a geometry problem, suggesting that it was liable to hallucinations.

Advanced voice mode with vision was delayed several times – allegedly Partly because OpenAI announced the feature long before it was ready for production. In April, OpenAI promised that Advanced Voice Mode can be available to users “inside a number of weeks.” Months later, the corporate said it needed more time.

When Advanced Voice Mode finally rolled out to some ChatGPT users in early fall, it was missing the visual analytics component. Ahead of Thursday's launch, OpenAI has turned its attention to bringing the voice-only Advanced Voice Mode experience to additional platforms Users within the EU.

Competitors like Google and Meta are working on similar features for his or her respective chatbot products. This week, Google made its real-time video analytics conversation AI feature Project Astra available to a bunch of “trusted testers” on Android.

In addition to Advance Voice Mode with Vision, OpenAI on Thursday introduced a festive “Santa Mode” that adds Santa's voice as a preset voice in ChatGPT. Users can find it by tapping or clicking the snowflake icon next to the prompt bar within the ChatGPT app.

LEAVE A REPLY

Please enter your comment!
Please enter your name here

Must Read