OpenAI releases ChatGPT’s hyper-realistic voice to some paying users
OpenAI introduced GPT-4o's Advanced Voice Mode to a limited group of ChatGPT Plus users with hyper-realistic audio responses.
OpenAI began deploying GPT-4o’s Advanced Voice Mode to select ChatGPT Plus users on Tuesday, allowing them to experience hyper-realistic audio responses. Initially accessible to a small group, the full rollout is planned for fall 2024, but video and screensharing features are not yet included and will launch later.
In May, the feature stunned audiences with its real-human resemblance, notably sounding like actress Scarlett Johansson. Despite OpenAI denying the use of Johansson’s voice, legal complications led to the removal of the voice from the demo, delaying the Advanced Voice Mode’s release to enhance safety measures.
The new voice mode uses GPT-4o without auxiliary models, promising lower latency with emotional intonation recognition. To avoid deepfake issues, OpenAI has limited the feature to four preset voices and included new filters to block unauthorized requests, addressing potential legal concerns surrounding voice impersonation and copyrighted audio.