ChatGPT Can Now Understand and Speak Multiple Languages and Listen to You
ChatGPT is entering the next phase of its development, as OpenAI has announced this week that the popular AI chatbot is getting new features, including speech and hearing prompts.
These features provide a new, more intuitive user interface by enabling voice chat or showing ChatGPT what you’re talking about, the company said in a statement.
Altman even shared his excitement about the new updates via this post:
voice mode and vision for chatgpt! really worth a try. https://t.co/g8uA4QxXMb— Sam Altman (@sama) September 25, 2023
OpenAI will offer voice and images on ChatGPT to Plus subscribers and business users over the next couple of weeks. It also noted that ChatGPT’s voice feature is available on both iOS and Android devices, and you have to manually enable the feature from your phone’s settings. OpenAI also confirmed that ChatGPT’s image feature works on all platforms.
So what has OpenAI done to make ChatGPT speak and hear prompts?
It says new features like voice are based on a new text-to-speech model that can deliver a voice that sounds very human. All it needs is text and a few seconds of sample speech to make the magic work.
“We worked with professional voice actors to create each voice. We also use Whisper, our open-source speech recognition system, to transcribe your spoken words into text,” OpenAI said.
For images, ChatGPT uses the power of GPT versions 3.5 and 4 to practice and master a variety of images with text as well.
OpenAI is obviously excited, but equally concerned about the potential abuse of these features. “These features also introduce new risks, such as the ability for malicious actors to impersonate public figures or commit fraud,” the company noted.
“We have also taken technical measures to significantly limit ChatGPT’s ability to analyze and make direct statements about people because ChatGPT is not always accurate and these systems must respect individuals’ privacy,” the company said.