After numerous delays and postponements, OpenAI has finally launched its new voice mode for ChatGPT. Currently, this feature is being rolled out to a select group of ChatGPT Plus Subscribers. Utilizing the latest GPT-4o AI model, this advanced voice mode provides real-time responses, a natural-sounding voice, and the ability to detect user emotions.
Originally scheduled for release as as alpha version in June, the rollout was pushed back by a month. OpenAI noted that the new voice mode will enable users to interrupt the AI at any time and facilitate more natural interactions through voice modulations. The company has also posted a brief video on X, demonstrating how to activate the feature once it becomes available.
Also Read: Sam Altman Calls For Name Change For ChatGPT Following New Model Launch
OpenAI’s new Voice Mode: How does it work
The advanced Voice Mode utilizes a sophisticated AI model that converts the user’s voice input into text through speech recognition technology. This text is then processed by ChatGPT’s language model to generate an appropriate response, which is subsequently converted into a female voice using a text-to-speech model.
Currently, this advanced voice mode is being tested with a limited group of ChatGPT Plus users. OpenAI stated that the selected users will receive an email with instructions and a message in their mobile app. OpenAI intends to gradually include more users and aims for all Plus subscribers to have access to voice mode by the end of the year.