The views expressed by contributors are their own and not the view of The Hill

America may not be ready for the looming tsunami of ‘deep fakes’

Open AI’s recent release of the DALL-E 2 text-to-image generator and Meta’s subsequent announcement of its “Make a Video” tool could erode barriers to creating “deep fakes,” synthetic images or video content created through artificial intelligence.

These new products, along with similar tools being developed by other companies, may provide significant advances in promoting the next wave of digital creators. At the same time, they could be exploited to create deep fakes that could have unintended economic, social and geopolitical consequences.

Images and videos have been falsified since the first photographs were taken. After the Civil War, “spirit photographers” claimed to be able to capture photos of deceased loved ones (including President Abraham Lincoln), ultimately to be revealed as frauds by P.T. Barnum. But deep fakes are fundamentally different. Their realism, and the scale and ease with which they can be produced, make them incredibly potent disinformation tools.

Americans may not be ready for this tsunami wave of deep fakes. In our recent research, subjects struggled to distinguish between deep fakes and authentic videos. When we randomly assigned a set of deep fake and authentic videos to more than 2,000 individuals and asked them to pick the deep fake, our test subjects were wrong over one-third of the time. Perhaps unsurprisingly given the social media savviness of American youth, middle school students outperformed adults, including the educators who might be responsible for helping them learn key skills to avoid online misinformation. Even computer science students at a top U.S. engineering university were susceptible: They were unable to sort out deep fakes from authentic videos more than 20 percent of the time.    

This could be a huge vulnerability that could be exploited to spread disinformation. During the initial stages of the Russian invasion of Ukraine, a deep fake of Ukrainian President Volodymyr Zelensky urging Ukrainian forces to surrender was circulated. While this video was easily debunked, it’s easy to see how this could have had geopolitical consequences. And other deep fakes have had real-world consequences, like the AI-generated voice deep fake that recently scammed a UK-based CEO out of nearly $250,000 of his company’s money.

Regulating deep fakes could be a dilemma for U.S. policymakers. Several states, including Texas, California and Virginia, have passed statutes prohibiting the release of deep fakes related to elections or child pornography. But regulations concerning other applications of deep fakes are not being actively considered. There have been discussions in some legal journals about whether First Amendment protections should cover deep fakes, but the question has not been resolved. 

And it’s too late for policymakers in the U.S. or the EU to stomp the brakes on the development and commercialization of deep fake technologies. Even if firms were somehow banned from advancing their work in this space, deep fake technology is already out in the wild. Novice programmers can readily use existing technologies to create convincing deep fakes, and the West’s geopolitical foes have been making steady advancements on their own. 

Technology could provide some assistance in detecting deep fakes, but a technological arms race is already ongoing between deep fake generators and automatic detectors — and the deep fake generators are winning. Our research showed that individuals might be better able to identify deep fakes from authentic videos if they understand the social context of a video. Still, America might need a concerted effort to advance scalable policy, social and technical solutions. Otherwise, the public could find itself drowning in a flood of deep fakes, with potentially disastrous consequences.

Jared Mondschein is a physical scientist at the nonprofit, nonpartisan RAND Corporation who researches AI policy and misinformation.

Christopher Doss is a policy researcher at RAND who specializes in fielding causal and descriptive studies at the intersection of early childhood education and education technologies.

Conrad Tucker is the Arthur Hamerschlag Career Development Professor of Mechanical Engineering and holds courtesy faculty appointments in machine learning, robotics, and biomedical engineering at Carnegie Mellon University. His research focuses on the design and optimization of systems through the acquisition, integration, and mining of large scale, disparate data.

Valerie Fitton-Kane is the vice president, Development, Partnerships, and Strategy, at Challenger Center, a not-for-profit leader in science, technology, engineering, and math (STEM) education, providing more than 250,000 students annually with experiential education programs that engage students in hands-on learning opportunities.

Lance Bush is president & CEO at Challenger Center, which was founded in 1986 by the STS-51L Challenger shuttle crew.

The opinions expressed in this publication are those of the authors. They do not purport to reflect the opinions or views of Carnegie Mellon University, RAND Corporation or Challenger Center.