This Tuesday, ChatGPT, a chatbot and virtual assistant developed by OpenAI announced that they are rolling out the advanced Voice Mode to some paid users.
Advanced Voice Mode for ChatGPT Roll Out
First time, OpenAI announced this feature in its Spring event during May this year.
Powered by the latest GPT-4o artificial intelligence (AI) model,
When it comes to OpenAI’s advanced Voice Mode, this offers features such as real-time responses, natural voice, and the capability to sense the user’s emotions.
The company has announced to offer this feature to all ChatGPT Plus users by fall this year.
Although, they haven’t provided any insight on when the video and screen sharing features, which were also demoed at the event, will be released.
Earlier OpenAI made an announcement regarding the rolling out of the advanced voice capabilities of ChatGPT in a post on X (formerly known as Twitter).
How Does This New Voice Mode Helps?
This new Voice Mode will allow users to interrupt the AI chatbot at any time and offer more natural interaction with voice modulations, according to the company.
They have also released a short video which highlighted how to turn on the feature once it becomes active.
According to this, the select group of ChatGPT Plus users will see an invite notification at the bottom of the screen prompting them to try the advanced Voice Mode after opening the app.
Once you tap on it, it will take the users to a new page with the title “You’re invited to try the advanced Voice Mode” and a button to activate the feature.
Presently, this feature is available to a small group of Plus users.
So far, the company did not specify any eligibility criteria.
The company has dubbed it as alpha roll out and this feature is powered by OpenAI’s latest flagship large language model (LLM), GPT-4o.
“Since we first demoed advanced Voice Mode, we’ve been working to reinforce the safety and quality of voice conversations as we prepare to bring this frontier technology to millions of people,” the AI firm said, while explaining the reason behind the delay.
OpenAI claims that GPT-4o’s voice capabilities have been tested with more than 100 external red teamers across 45 languages.
Here mentioned Red teamers are cybersecurity professionals who are tasked with testing a product or organization’s security by simulating cyberattacks and jailbreak attempts.
The whole goal of this process is to expose the vulnerabilities in the system before it goes live.
Presently, the user will only access four preset voices after the feature is rolled out to their account.
It is noteworthy here that Sky, the controversial voice which allegedly bore close similarities with actor Scarlett Johannson, is yet to be added back to ChatGPT.