A more human-like chatbot experience is on the horizon: OpenAI's New Voice Feature Goes Beyond Text
OpenAI has begun rolling out its Advanced Voice Mode for ChatGPT, allowing select users to engage in more natural, real-time conversations with the chatbot. This feature, currently in a limited testing phase, offers a dynamic shift from typical chatbot interactions, bringing a more conversational, human-like experience to the forefront.
Who’s Getting Access First?
Currently, Advanced Voice Mode is available only to a small group of ChatGPT Plus subscribers, who pay $20 per month for enhanced features such as faster responses and early access to new capabilities. Selection for this feature is based on various factors, although OpenAI hasn't disclosed the exact criteria. Long-term Plus users and active participants in other test phases like SearchGPT seem to have an edge, but the selection process remains varied.
According to an OpenAI message shared by a user on X (formerly Twitter), the alpha testing phase officially began on September 24, 2024, and the feature is expected to expand to more users throughout autumn. Free-tier ChatGPT users will have to wait, as their access depends on feedback from this current testing phase.
Enhancing Conversations: What’s New?
The standout benefit of Advanced Voice Mode is its ability to mimic real human speech patterns, making interactions feel more fluid and natural. Users can switch seamlessly between voice and text, offering flexibility depending on their needs—whether they’re asking quick questions or engaging in more extended discussions.
The voice feature also includes nuanced tone and inflection, which adapts to the context of the conversation, creating a smoother and more engaging interaction.
Compatibility and Setup
Advanced Voice Mode is designed for iOS and Android devices. iPhone users will need iOS 16.4 or later, and Android users should ensure they have the latest ChatGPT app version. Once updated, enabling voice mode is as simple as granting microphone permissions.
For optimal use, OpenAI suggests activating the feature in quieter settings or using headphones, as background noise can impact voice recognition. iPhone users can enable "Voice Isolation" to minimize distractions in noisier environments.
What’s Next for Advanced Voice Mode?
OpenAI plans to expand Advanced Voice Mode's capabilities by incorporating video and screen-sharing features in future updates. However, the immediate focus is on perfecting the voice interaction experience based on user feedback. As improvements roll out, more users can expect to be invited to test the feature, and a broader release may be on the horizon.