OpenAI has finally released the real-time video capabilities for ChatGPT that it demoed nearly seven months ago.
On Thursday during a livestream, the company said that Advanced Voice Mode, its human-like conversational feature for ChatGPT, is getting vision. Using the ChatGPT app, users subscribed to ChatGPT Plus, Team, or Pro can point their phones at objects and have ChatGPT respond in near real time.
Advanced Voice Mode with vision can also understand what’s on a device’s screen via screen sharing. It can explain various settings menus, for example, or give suggestions on a math problem.
To access Advanced Voice Mode with vision, tap the voice icon next to the ChatGPT chat bar, then tap the video icon on the bottom left, which will start video. To screen-share, tap the three-dot menu and select “Share Screen.”
The rollout of Advanced Voice Mode with vision will start Thursday, OpenAI says, and wrap up in the next week. But not all users will get access. OpenAI says that ChatGPT Enterprise and Edu subscribers won’t get the feature until January, and that it has no timeline for ChatGPT users in the EU, Switzerland, Iceland, Norway, or Liechtenstein.
In a recent demo on CBS News’ “60 Minutes,” OpenAI President Greg Brockman had Advanced Voice Mode with vision quiz Anderson Cooper on his anatomy skills. As Cooper drew body parts on a blackboard, ChatGPT could “understand” what he was drawing.

“The location is spot on,” ChatGPT said. “The brain is right there in the head. As for the shape, it’s a good start. The brain is more of an oval.”
Disrupt 2026: The tech ecosystem, all in one room
Your next round. Your next hire. Your next breakout opportunity. Find it at TechCrunch Disrupt 2026, where 10,000+ founders, investors, and tech leaders gather for three days of 250+ tactical sessions, powerful introductions, and market-defining innovation. Register now to save up to $400.
Save up to $300 or 30% to TechCrunch Founder Summit
1,000+ founders and investors come together at TechCrunch Founder Summit 2026 for a full day focused on growth, execution, and real-world scaling. Learn from founders and investors who have shaped the industry. Connect with peers navigating similar growth stages. Walk away with tactics you can apply immediately
Offer ends March 13.
In that same demo, Advanced Voice Mode with vision made a mistake on a geometry problem, however, suggesting that it’s prone to hallucinating.
Advanced Voice Mode with vision has been delayed multiple times — reportedly in part because OpenAI announced the feature far before it was production-ready. In April, OpenAI promised that Advanced Voice Mode would roll out to users “within a few weeks.” Months later, the company said it needed more time.
When Advanced Voice Mode finally arrived in early fall for some ChatGPT users, it lacked the visual analysis component. In the lead-up to Thursday’s launch, OpenAI has focused its attention on bringing the voice-only Advanced Voice Mode experience to additional platforms and users in the EU.
Rivals like Google and Meta are working on similar capabilities for their respective chatbot products. This week, Google made its real-time, video-analyzing conversational AI feature, Project Astra, available to a group of “trusted testers” on Android.
In addition to Advance Voice Mode with vision, OpenAI on Thursday launched a festive “Santa Mode,” which adds Santa’s voice as a preset voice in ChatGPT. Users can find it by tapping or clicking the snowflake icon in the ChatGPT app next to the prompt bar.
