Connect with us

Unique News

As Deepfakes Get Faker, What Does This Mean for Our Future? – Adweek

As Deepfakes Get Faker, What Does This Mean for Our Future? – Adweek


As Deepfakes Get Faker, What Does This Mean for Our Future? – Adweek

Editor’s note: Industry consultant Shelly Palmer is taking his popular newsletter and turning it into an Adweek article once per week in an ongoing column titled “Think About This.”

Late last year, I wrote about deepfakes and how, for about 200,000 years, we have relied on our eyes and ears to separate truth from lies and fact from fiction. Even if we ignore the rise of fake news, technology is on the verge of making it impossible to know if what we are seeing and hearing is real or fake.

It’s been less than a year since I wrote that article and the technology has advanced in ways that make last year’s fakes look primitive. Here’s a roundup of the advancements and uses in the past year.

Already illegal: deepnudes

Revenge porn has been outlawed in many jurisdictions around the globe. In July, Virginia became one of the first states to outlaw the sharing of computer-generated pornography (aka deepfakes). It did so by amending an existing law that criminalizes revenge porn, making clear that the category now includes “falsely-created” material.

Scary: DNC deepfake

The Democratic Party deepfaked its own chairman to highlight 2020 concerns. In early August at DEF CON, one of the world’s biggest conventions for hackers, attendees were told DNC chair Tom Perez couldn’t make it for the DNC’s presentation. In lieu of that, he “Skyped in” and chatted. Except he didn’t. Do you know what Perez’s voice sounds like? Neither did those in attendance.

Bob Lord, the DNC’s chief security officer, led the exercise with the help of experts in the field of AI as a way to demonstrate the art-of-the-possible. This is what’s coming leading up to the 2020 election. The video speaks for itself. Please watch it.

Fun: fixing The Lion King

It’s important to remember that not all deepfakes are malicious. After Disney purists didn’t like the computer graphics in this summer’s release of The Lion King, they “fixed” the trailer by melding the original animated version’s character faces onto the new version’s animals. The resulting footage is so well done that it takes a second to realize what you’re looking at.

Somewhere in between: FaceApp

FaceApp, an app on Android and iOS that uses AI to “generate highly realistic transformations of faces in photographs,” burst on the scene this summer. Basically, FaceApp puts the power of deepfakes into your hand, making it a snap to make a person in a photo look older, younger or change genders.

Aside from the mild controversy—including the discovery that every photo uploaded to the app was sent to FaceApp’s cloud servers—the tech here is remarkable. FaceApp has been offering its photo effects since early 2017, but the results have gotten much better in the years since it launched.

While FaceApp has limited features, it’s one of the easiest and most popular deepfake creators available to the masses. Its speed and impressive results make it fun—and also dangerous.

What does this all mean?

The technology behind deepfakes was already easy to find and use last fall, but it’s even more so today. Imagine where it’ll be next summer when the 2020 election is four months away. Did Elizabeth Warren really say that? Was that footage of Bernie Sanders in that attack ad real, or was it a deepfake? As political advertisements already twist candidates’ words and manipulate the truth for the perfect soundbite, can you believe anything you hear when it can all be manufactured on any laptop?

It gets worse

While news cycle after news cycle covered allegations of possible social media manipulation prior to the 2016 election, is anyone ready for what comes next? Let’s quickly imagine how an AI-assisted system will exacerbate the problem.

Source link

Continue Reading
You may also like...
Click to comment

Leave a Reply

Your email address will not be published. Required fields are marked *


To Top
error: Content is protected !!