news

OpenAI launches GPT-4 speech model, which will be available to all paying users this fall

2024-07-31

한어Русский языкEnglishFrançaisIndonesianSanskrit日本語DeutschPortuguêsΕλληνικάespañolItalianoSuomalainenLatina

OpenAI, an American artificial intelligence research company, recently announced that it will start rolling out GPT-4's voice mode to some ChatGPT Plus users. This advanced voice mode can provide more natural real-time conversations, and users can interrupt the conversation at any time, and GPT-4 can sense and respond to the user's emotions. The launch of this feature means another important breakthrough for OpenAI in voice assistant technology, especially the two recognized technical difficulties of real-time response and interruptible conversations.

OpenAI launched a new version of its large model GPT-4 in May this year, and also demonstrated its speech mode. It was originally planned to gradually open the speech mode to users at the end of June, but it was finally decided to postpone the release to July. According to OpenAI's plan, the speech mode will be open to all ChatGPT Plus users this fall.

Currently, GPT-4's voice model provides four preset voices, Juniper, Breeze, Cove, and Ember, which were created by OpenAI in collaboration with paid voice actors.