Skip to main content
  1. Home
  2. Emerging Tech
  3. Computing
  4. News

AI can now duplicate anyone's voice based on just one minute of training

Add as a preferred source on Google

Do you remember the cool Mission Impossible tech that lets Tom Cruise’s character Ethan Hunt mimic the voice of other characters using some nifty speech synthesis technology?

Well, a Montreal-based startup called Lyrebird (named after the sound-imitating bird) just invented it for real.

Recommended Videos

“We are developing new speech synthesis technologies which, among other features, allow us to copy the voice of someone with very little data,” Alexandre de Brebisson, one of the PhD students who developed the deep-learning tech behind the project. “Our experiments show that one minute of audio already contains a lot of the DNA of a human voice. We are able to learn a new voice with as little data because our model is able to capture similarities between the new voice and all the voices it already knows. Our models understand the underlying variables that make [one] voice different from another.”

Since the tech was shown off this week, de Brebisson said his team have received dozens of different suggested use-cases by email, some containing applications they’d thought of, and others containing ones that they hadn’t.

Some companies, for example, are interested in letting their users choose to have audio books read in the voice of either famous people or family members. The same is true of medical companies, which could allow people with voice disabilities to train their synthetic voices to sound like themselves, if recorded samples of their speaking voices exist. Another interesting idea is for video game companies to offer the ability for in-game characters to speak with the voice of the human player.

There are plenty more exciting opportunities, which have led to 10,000 people already signing up to be informed of the forthcoming beta version. “We will then add features over time, such as letting companies design a unique voice tailored for their needs, and control the emotion of the [voice] generation,” de Brebisson continued.

While it doesn’t sound perfect yet, it’s not hard to imagine how this might sound in just a few years. Combined with technology such as software for making convincing edits to the moving lips of a person who is speaking, “fake news” circa 2025 should certainly be a whole lot of fun.

Right?

Luke Dormehl
I'm a UK-based tech writer covering Cool Tech at Digital Trends. I've also written for Fast Company, Wired, the Guardian…
Robots just ran the Beijing half-marathon faster than the world record holder
humanoid robot running a marathon

A humanoid robot just ran a half-marathon faster than the world record holder. It might not seem impressive at first, but considering last year, the fastest robot at Beijing's humanoid robot half-marathon finished in two hours and 40 minutes, this is a huge achievement. 

As reported by the Associated Press, the winning robot at this year's Beijing half-marathon crossed the finish line in 50 minutes and 26 seconds, comfortably beating the human world record of 57 minutes recently set by Jacob Kiplimo. 

Read more
As if the plate wasn’t already full, AI is about to worsen the global e-waste crisis
New report highlights a rising environmental concern
Stack of graphics cards and motherboards in a landfill site e-waste

AI is already changing how the world works, but it’s also quietly making one of our biggest environmental problems even worse. And no, this isn’t about energy consumption this time. It’s about the hardware. Because every smarter AI model comes with a physical cost.

AI is about to supercharge the e-waste problem

Read more
Smart glasses are finding a surprise niche — Korean drama and theater shows
Urban, Night Life, Person

Every year, millions of people follow Korean content without speaking a word of the language. They stream shows with subtitles, read translated lyrics, and find workarounds. But live theater has always been a different problem — you can't pause or rewind it. That's the problem: a Korean startup thinks it's cracked, and Yuroy Wang was one of the first to try it. The 22-year-old Taipei retail worker is a K-pop fan who loves Korean culture but doesn't speak the language. When he went to see "The Second Chance Convenience Store," a touring play based on a Korean novel that was a bestseller in Taiwan, he expected supertitles. What he got instead was a pair of chunky black-framed AI-powered glasses sitting on his nose, translating the dialogue in real time directly on the lenses. "As soon as I found out they were available, I couldn't wait to try them," he said. Wang is part of a growing audience discovering that smart glasses, a category of tech that has struggled to find mainstream purpose for years, might have just found their calling in the most unexpected of places: live Korean theater.

How do the glasses work?

Read more