What version
Users report missing voice features as OpenAI delays advanced Voice Mode rollout beyond initial June target.
OpenAI's highly anticipated advanced Voice Mode for ChatGPT-4o, first showcased in May 2024, has missed its initial target for a limited alpha release in June. This delay has led to widespread confusion among users, who are seeing references to the feature in demos and documentation but cannot access it in their own accounts. The advanced Voice Mode promised real-time, conversational audio interactions with emotional range and the ability to interrupt the AI—a significant leap beyond the existing Read Aloud feature, which simply vocalizes text responses.
The company attributes the holdup to the need for further internal safety and reliability evaluations, particularly around content moderation and latency. OpenAI has stated the feature will first launch for a small group of ChatGPT Plus subscribers in an alpha phase, with a broader rollout planned for Fall 2024. In the interim, users only have access to the standard Read Aloud function. This gap between demo hype and product availability highlights the challenges of deploying complex, real-time AI systems at scale while managing user expectations.
- Advanced Voice Mode for ChatGPT-4o was demoed in May with a planned June alpha launch that has been delayed.
- The feature enables real-time, expressive conversations and is distinct from the basic text-to-speech 'Read Aloud' tool.
- OpenAI cites safety and scalability testing for the delay, with a phased rollout now expected starting Fall 2024.
Why It Matters
The delay underscores the difficulty in shipping complex multimodal AI features, affecting developer timelines and user trust.