I never said that! High-tech deception of 'deepfake' videos

Hey, did my congressman really say that? Is that really President Donald Trump on that video, or am I being duped?

New technology on the internet lets anyone make videos of real people appearing to say things they've never said. Republicans and Democrats predict this high-tech way of putting words in someone's mouth will become the latest weapon in disinformation wars against the United States and other Western democracies.

We're not talking about lip-syncing videos. This technology uses facial mapping and artificial intelligence to produce videos that appear so genuine it's hard to spot the phonies. Lawmakers and intelligence officials worry that the bogus videos – called deepfakes – could be used to threaten national security or interfere in elections.

So far, that hasn't happened, but experts say it's not a question of if, but when.

"I expect that here in the United States we will start to see this content in the upcoming midterms and national election two years from now," said Hany Farid, a digital forensics expert at Dartmouth College in Hanover, New Hampshire. "The technology, of course, knows no borders, so I expect the impact to ripple around the globe."

When an average person can create a realistic fake video of the president saying anything they want, Farid said, "we have entered a new world where it is going to be difficult to know how to believe what we see."

The reverse is a concern, too. People may dismiss as fake genuine footage, say of a real atrocity, to score political points.

Realizing the implications of the technology, the U.S. Defense Advanced Research Projects Agency is already two years into a four-your program to develop technologies that can detect fake images and videos. Right now, it takes extensive analysis to identify phony videos. It's unclear if new ways to authenticate images or detect fakes will keep pace with deepfake technology.

Deepfakes are so named because they utilize deep learning, a form of artificial intelligence. They are made by feeding a computer an algorithm, or set of instructions, lots of images and audio of a certain person. The computer program learns how to mimic the person's facial expressions, mannerisms, voice and inflections. If you have enough video and audio of someone, you can combine a fake video of the person with a fake audio and get them to say anything you want.

So far, deepfakes have mostly been used to smear celebrities or as gags, but it's easy to foresee a nation state using them for nefarious activities against the U.S., said Sen. Marco Rubio, R-Fla., one of several members of the Senate intelligence committee who are expressing concern about deepfakes.

A foreign intelligence agency could use the technology to produce a fake video of an American politician using a racial epithet or taking a bribe, Rubio says. They could use a fake video of a U.S. soldier massacring civilians overseas, or one of a U.S. official supposedly admitting a secret plan to carry out a conspiracy. Imagine a fake video of a U.S. leader – or an official from North Korea or Iran – warning the United States of an impending disaster.

"It's a weapon that could be used – timed appropriately and placed appropriately – in the same way fake news is used, except in a video form, which could create real chaos and instability on the eve of an election or a major decision of any sort," Rubio told The Associated Press.

Deepfake technology still has a few hitches. For instance, people's blinking in fake videos may appear unnatural. But the technology is improving.

"Within a year or two, it's going to be really hard for a person to distinguish between a real video and a fake video," said Andrew Grotto, an international security fellow at the Center for International Security and Cooperation at Stanford University in California.

"This technology, I think, will be irresistible for nation states to use in disinformation campaigns to manipulate public opinion, deceive populations and undermine confidence in our institutions," Grotto said.

He called for government leaders and politicians to clearly say it has no place in civilized political debate.

X
Sitelerimizde reklam ve pazarlama faaliyetlerinin yürütülmesi amaçları ile çerezler kullanılmaktadır.

Bu çerezler, kullanıcıların tarayıcı ve cihazlarını tanımlayarak çalışır.

İnternet sitemizin düzgün çalışması, kişiselleştirilmiş reklam deneyimi, internet sitemizi optimize edebilmemiz, ziyaret tercihlerinizi hatırlayabilmemiz için veri politikasındaki amaçlarla sınırlı ve mevzuata uygun şekilde çerez konumlandırmaktayız.

Bu çerezlere izin vermeniz halinde sizlere özel kişiselleştirilmiş reklamlar sunabilir, sayfalarımızda sizlere daha iyi reklam deneyimi yaşatabiliriz. Bunu yaparken amacımızın size daha iyi reklam bir deneyimi sunmak olduğunu ve sizlere en iyi içerikleri sunabilmek adına elimizden gelen çabayı gösterdiğimizi ve bu noktada, reklamların maliyetlerimizi karşılamak noktasında tek gelir kalemimiz olduğunu sizlere hatırlatmak isteriz.