Skip to main content
  1. Home
  2. Emerging Tech
  3. Computing
  4. Features

A.I.-generated text is supercharging fake news. This is how we fight back

Add as a preferred source on Google
How Deep Fakes Will Make Fake News Worse - The Deets

Last month, an A.I. startup backed by sometimes A.I. alarmist Elon Musk announced a new artificial intelligence they claimed was too dangerous to release to the public. While “only” a text generator, OpenAI’s GPT2 was reportedly capable of generating text so freakishly humanlike that it could convince people that it was, in fact, written by a real flesh and blood human being.

Recommended Videos

To use GPT2, a user would only have to feed it the start of a document, before the algorithm would take over and complete it in a highly convincing manner. For instance, give it the opening paragraphs of a newspaper story and it would manufacture “quotes” and assorted other details.

Such tools are becoming increasingly common in the world of A.I. — and the world of fake news, too. The combination of machine intelligence and, perhaps, the distinctly human unintelligence that allows disinformation to spread could prove a dangerous mix.

Fortunately, a new A.I. developed by researchers at MIT, IBM’s Watson A.I. Lab and Harvard University is here to help. And just like a Terminator designed to hunt other Terminators, this one — called GLTR — is uniquely qualified to spot bot impostors.

Fighting the good fight

As its creators explain in a blog post, text generation tools like GPT2 open up “paths for malicious actors to … generate fake reviews, comments or news articles to influence the public opinion. To prevent this from happening, we need to develop forensic techniques to detect automatically generated text.”

GLTR takes the same models that are are used as the basis for fake text generation by GPT2. By looking at a piece of text, and then predicting which words the algorithm would likely have picked to follow one another, it can give a verdict on whether it thinks it was written by a machine. The tool is available for users to try online. (If anyone has ever told you that your own writing is too machine-like, this might be your chance to prove them wrong!)

 

GPT-2 generates synthetic text samples in response to the model being primed with an arbitrary input. The model is chameleon-like—it adapts to the style and content of the conditioning text. This allows the user to generate realistic and coherent continuations about a topic of their choosing. OpenAI

Until now, it’s been relatively easy for humans to pick out writing generated by machines — usually because it is overly formulaic or, in creative writing, makes little to no sense. That’s fast changing, though, and the creators of GLTR think that tools such as this will therefore become more necessary.

“We believe that machines and humans excel at detecting fundamentally different aspects of generated text,” Sebastian Gehrmann, a Ph.D. candidate in Computer Science at Harvard, told Digital Trends. “Machine learning algorithms are great at picking up statistical patterns such as the ones we see in GLTR. However, at the moment machines do not actually understand the content of a text. That means that algorithms could be fooled by completely nonsensical text, as long as the patterns match the detection. Humans, on the other hand, can easily tell when a text does not make any sense, but cannot detect the same patterns we show in GLTR.”

“Imagine getting emails or reading news, and a browser plugin tells you for the current text how likely it was produced by model X or model Y.”

Hendrik Strobelt, a data scientist at IBM Research, told us that figuring out whether a piece of text comes from a human origin will become more of a pressing issue. “[Our current] visual tool might not be the solution to that, but it might help to create algorithms that work like spam detection algorithms,” he said. “Imagine getting emails or reading news, and a browser plugin tells you for the current text how likely it was produced by model X or model Y.”

A cat and mouse game

Similar games of one upmanship — in which A.I. tools are used to spot fakes created by others A.I.s — are taking place across the tech industry. This is particularly true when it comes to fake news. For example, “deepfakes” have caused plenty of alarm with their promise of being able to realistically superimpose one person’s head onto another’s body.

To help counter deepfakes, researchers from Germany’s Technical University of Munich have developed an algorithm called XceptionNet that’s designed to quickly spot faked videos posted online. Speaking with Digital Trends last year, one of the brains behind XceptionNet suggested a similar approach involving a possible browser plugin that runs the algorithm continuously in the background.

It seems likely that others are working on solutions for spotting the A.I. behind other forms of machine-masquerading-as-humans, such as Google’s Duplex voice calling tech or the spate of artificial intelligences capable of accurately mimicking celebrity voices and making them say anything the user wants.

Image used with permission by copyright holder

This kind of cat-and-mouse game will be of no great shock to anyone who has followed the world of hacking. Hackers spot vulnerabilities in systems and exploit them, then somebody notices and patches the hole, leaving hackers to move onto the next vulnerability. In this case, however, the escalation involves cutting edge artificial intelligence.

“In the future, we will see increasingly common [use and] abuse of algorithmically generated text,” Gehrmann continued. “In only a few years, algorithms could potentially be used to influence the public opinion on products, movies, personalities, or politics on a larger and larger scale. Therefore, tools to detect fake content will become more and more relevant for real-world use. As researchers, we see it as our goal to develop detection methods at a faster rate than the generation methods to combat and extinguish this abuse.”

Now we just have to hope that the good guys can work harder and faster than the bad ones. Unfortunately, if history has taught us anything it’s that there’s guarantee that this will be the case. Keep your fingers crossed that it is!

Luke Dormehl
I'm a UK-based tech writer covering Cool Tech at Digital Trends. I've also written for Fast Company, Wired, the Guardian…
Robots just ran the Beijing half-marathon faster than the world record holder
humanoid robot running a marathon

A humanoid robot just ran a half-marathon faster than the world record holder. It might not seem impressive at first, but considering last year, the fastest robot at Beijing's humanoid robot half-marathon finished in two hours and 40 minutes, this is a huge achievement. 

As reported by the Associated Press, the winning robot at this year's Beijing half-marathon crossed the finish line in 50 minutes and 26 seconds, comfortably beating the human world record of 57 minutes recently set by Jacob Kiplimo. 

Read more
As if the plate wasn’t already full, AI is about to worsen the global e-waste crisis
New report highlights a rising environmental concern
Stack of graphics cards and motherboards in a landfill site e-waste

AI is already changing how the world works, but it’s also quietly making one of our biggest environmental problems even worse. And no, this isn’t about energy consumption this time. It’s about the hardware. Because every smarter AI model comes with a physical cost.

AI is about to supercharge the e-waste problem

Read more
Smart glasses are finding a surprise niche — Korean drama and theater shows
Urban, Night Life, Person

Every year, millions of people follow Korean content without speaking a word of the language. They stream shows with subtitles, read translated lyrics, and find workarounds. But live theater has always been a different problem — you can't pause or rewind it. That's the problem: a Korean startup thinks it's cracked, and Yuroy Wang was one of the first to try it. The 22-year-old Taipei retail worker is a K-pop fan who loves Korean culture but doesn't speak the language. When he went to see "The Second Chance Convenience Store," a touring play based on a Korean novel that was a bestseller in Taiwan, he expected supertitles. What he got instead was a pair of chunky black-framed AI-powered glasses sitting on his nose, translating the dialogue in real time directly on the lenses. "As soon as I found out they were available, I couldn't wait to try them," he said. Wang is part of a growing audience discovering that smart glasses, a category of tech that has struggled to find mainstream purpose for years, might have just found their calling in the most unexpected of places: live Korean theater.

How do the glasses work?

Read more