OpenAI introduces voice-cloning tool

OpenAI says ways to verify people consented to having their voices immitated by artificial intelligence and to automatically detect audio deep fakes involving prominent people should be built in widely deployed 'synthetic voice' tools © SEBASTIEN BOZON / AFP

OpenAI on Friday unveiled a voice-cloning tool that it intends to tightly control until safeguards are implemented to prevent the spread of audio forgeries designed to deceive listeners.

A model called “Voice Engine” can essentially duplicate someone’s speech based on a 15-second audio sample, according to an OpenAI blog post sharing results of a small-scale test of the tool.

“We recognize that generating speech that resembles people’s voices has serious risks, which are especially top of mind in an election year,” the San Francisco-based company said.

“We are engaging with U.S. and international partners from across government, media, entertainment, education, civil society and beyond to ensure we are incorporating their feedback as we build.”

Disinformation researchers fear rampant misuse of AI-powered applications in a pivotal election year thanks to proliferating voice cloning tools, which are cheap, easy to use and hard to trace.

Acknowledging these problems, OpenAI said it was “taking a cautious and informed approach to a broader release due to the potential for synthetic voice misuse.”

OpenAI said that partners testing Voice Engine agreed to rules including requiring explicit and informed consent of any person whose voice is duplicated using the tool.

It must also be made clear to audiences when voices they are hearing are AI generated, the company added.

“We have implemented a set of safety measures, including watermarking to trace the origin of any audio generated by Voice Engine, as well as proactive monitoring of how it’s being used,” OpenAI said.