The Voice Feature by OpenAI for ChatGPT Plus and Team Users
This week, OpenAI has released the Advanced Voice Mode feature for all ChatGPT Plus and Team subscribers, which enhances the dialogues experience with the running AI chatbot. New features that the activation brings are five new voice options and various enhancements to the experience of interacting with voices.
Introducing the High-Level Voice Mode for ChatGPT Plus
This week, OpenAI announced it is making the Advanced Voice Mode available to all ChatGPT Plus and Team subscribers. The feature, accessible in the ChatGPT app, permits interactions to be more fluid and occur in real-time. This summer, it instead launched as limited test available to a handful of users in July with this expanded availability targeted for the end of week.
But this feature is not yet live in several regions such as the European Union, UK, Switzerland, Norway, Iceland and Liechtenstein.
Five More Voice Styles and Better Accents
At the same time as the expansion, OpenAI has added five new voices—Arbor, Maple, Sol, Spruce and Vale—to take the total to nine. This individual personality provides different shades and moods of each voice helping users to engage with them as a better conversational experience. In addition to new accents in certain foreign languages, the company says that it has also implemented additional improvements to the general flow and speed of conversations.
Tailor Instructions and Boost Performance
This advanced Voice Mode also lets users create customized commands adds more versatility to the interactions. The new feature has led to an improvement in conversational latency, reducing response times down to 2.8 seconds for GPT-3. For GPT-4 models, the loading time is 5 and 5. That delay is caused by the data pipeline preprocessing audio to text, processing and converting it back to audio.
This issue is dealt with for the new GPT-4o model by fusing everything into a single neural network: text, vision, and audio — which results in lower latency and helps maintain natural conversation. Another benefit is that the chatbot can deal with interruptions, group conversations, and background noise in a more robust way as well.
Restrictions & Security Controls
Several restrictions are in place to prevent misuse of Advanced Voice Mode by OpenAI. This limits the feature to a pre-canned selection of voices created by professional voice actors, prevents impersonation and adds filters for requests around violent content or copyrighted material, like music and sound.
Aside from those improvements, though, not everything that was originally shown is there. It can’t do something like recognize and critique particular dance moves over a smartphone camera yet.
Increasing Popularity of Smart Voice Assistants
With a growing number of voice assistants being developed, the ease of user experience may be worth the paid download to some. Twenty-two percent of Gen Z users have a strong willingness to spend $10+ per month on a voice assistant that exhibits human-level intelligence and accuracy, reports PYMNTS Intelligence. This increased interest points to a large market demand for ever-more powerful voice capabilities in AI-powered applications, including but not limited to ChatGPT.
Wrap-Up: Added Voice Mode to ChatGPT — A great way forward
The OpenAI team further challenges the boundaries of AI interaction with Advanced Voice Mode. Five additional new voices, enhanced performance and personalized settings will provide a more diverse user experience. With the growing power of OpenAI voice, users can anticipate a host of other upcoming, as-yet-unheard-of capabilities in next-generation technologies powered by MTIDs.