Realistic Fake Videos Threaten Democracy

Disinformation and distrust online are set to take a turn for the worse. Rapid advances in deep-learning algorithms to synthesize video and audio content have made possible the production of “deep fakes”, highly realistic and difficult-to-detect depictions of real people doing or saying things they never said or did. 

As this technology spreads, the ability to produce bogus yet credible video and audio content will come within the reach of an ever-larger array of governments, non-state actors, and individuals. As a result, the ability to advance lies using hyper-realistic, fake evidence is poised for a great leap forward.

The array of potential harms that deep fakes could entail is stunning. A well-timed and thoughtfully scripted deep fake or series of deep fakes could tip an election, spark violence in a city primed for civil unrest, bolster insurgent narratives about an enemy’s supposed atrocities, or exacerbate political divisions in a society. 

The opportunities for the sabotage of rivals are legion, for example, sinking a trade deal by slipping to a foreign leader a deep fake purporting to reveal the insulting true beliefs or intentions of US officials.

The prospect of a comprehensive technical solution is limited for the time being, as are the options for legal or regulatory responses to deep fakes. A combination of technical, legislative, and personal solutions could help stem the problem.

Background: What Makes Deep Fakes Different?

The creation of false video and audio content is not new. Those with resources, like Hollywood studios or government entities, have long been able to make reasonably convincing fakes. The “appearance” of 1970s-vintage Peter Cushing and Carrie Fisher in Rogue One: A Star Wars Story is a recent example.

The looming era of deep fakes will be different, however, because the capacity to create hyper-realistic, difficult-to-debunk fake video and audio content will spread far and wide. Advances in machine learning are driving this change. 

Most notably, academic researchers have developed “generative adversarial networks” (GANs) that pit algorithms against one another to create synthetic data (i.e., the fake) that is nearly identical to its training data (i.e., real audio or video). Similar work is likely taking place in various classified settings, but the technology is developing at least partially in full public view with the involvement of commercial providers. 

Some degree of credible fakery is already within the reach of leading intelligence agencies, but in the coming age of deep fakes, anyone will be able to play the game at a dangerously high level. 

In such an environment, it would take little sophistication and resources to produce havoc. Not long from now, robust tools of this kind and for-hire services to implement them will be cheaply available to anyone.

The information-sharing environment is well suited to the spread of falsehoods. In the United States and many other countries, society already grapples with surging mis-information resulting from the declining influence of quality-controlled mass media and the growing significance of social media as a comparatively unfiltered, many-to-many news source. 

As of August 2017, two-thirds of Americans reported to Pew that they get their news at least in part from social media. This is fertile ground for circulating deep fake content. Indeed, the more salacious, the better.

Foreign Policy Implications

The arrival of deep fakes has frightening implications for foreign affairs and national security. They could be potent instruments of covert action campaigns and other forms of disinformation used in international relations and military operations, with potential for serious damage. 

The information operation against Qatar in 2017, which attributed pro-Iranian views to Qatar’s emir, illustrates how significant fraudulent content can be even without credible audio and video.

For example, a credible deep fake audio file could emerge purporting to be a recording of President Donald J. Trump speaking privately with Russian President Vladimir Putin during their last meeting in Helsinki, with Trump promising Putin that the United States would not defend certain North Atlantic Treaty Organisation (NATO) allies in the event of Russian subversion. 

Other examples could include deep fake videos depicting an Israeli soldier committing an atrocity against a Palestinian child, a European Commission official offering to end agricultural subsidies on the eve of an important trade negotiation, or a Rohingya leader advocating violence against security forces in Myanmar.

Democracy could suffer as well. The circulation of a plausible video clip depicting a candidate uttering despicable things twenty-four hours before an election could control the outcome. 

Short of that, deep fakes would allow for more effective disinformation operations similar to Russia’s efforts against the US presidential election in 2016. As the technology diffuses, a widening circle of non-state actors and individuals would be able to cause similar problems.

The Challenge of Limiting the Harms

There is no silver-bullet solution to this problem, and certainly no option of rolling back the technological progress that makes deep fakes possible. Worse, some of the most plausible responses carry significant costs of their own.

Ideally, this technology-driven problem could be addressed adequately through technological solutions. But though strong detection algorithms are emerging (including GAN-based methods), they are lagging behind the innovation found in the creation of deep fakes. 

Even if an effective detection method emerges, it will struggle to have broad impact unless the major content distribution platforms, including traditional and social media, adopt it as a screening or filtering mechanism. 

The same is true for potential solutions involving digital provenance: video or audio content can be watermarked at its creation, producing immutable metadata that marks location, time, and place and attests that the material was not tampered with. To have a broad effect, digital provenance solutions would need to be built into all the devices people use to create content, and traditional and social media would need to incorporate those solutions into their screening and filtering systems. 

However, there is little reason to expect convergence on a common standard for digital provenance, let alone to expect that such technology would be adopted in those ways.

Another option would be for Congress to intervene with regulatory legislation compelling the use of such technology, but that approach would entail a degree of market intervention unlike anything seen previously with respect to these platforms and devices. 

This option would also run the risk of stifling innovation due to the need to pick winners even while technologies and standards continue to evolve.

Legal and regulatory frameworks could play a role in mitigating the problem, but as with most technology-based solutions they will struggle to have broad effect, especially in the case of international relations. Existing laws already address some of the most malicious fakes; a number of criminal and tort statutes forbid the intentional distribution of false, harmful information. 

But these laws have limited reach. It is often challenging or impossible to identify the creator of a harmful deep fake, and they could be located outside the United States. Foreign actors creating deep fakes can be named and shamed, but the ongoing fallout from Russian election interference in 2016 illustrates the limits of that approach.

Another possibility is to pressure traditional and social media platforms to do more to identify and suppress deep fakes, a familiar proposition in today’s ongoing debate about disinformation and social media. 

Companies like Facebook are in a bottleneck position well-suited for preventing the broad distribution of harmful content. Facebook, and some other platforms, have responded to recent congressional pressure by showing serious interest in improving the quality of their filtering systems. Still, past performance suggests the need for a dose of skepticism in relation to such efforts.

Social media platforms have long been insulated from liability for distributing harmful content. Section 230 of the Communications Decency Act of 1996 broadly immunises online service providers in relation to harms caused by user-generated content, with only a few exceptions. 

Congress could give platforms stronger incentives to self-police by limiting that immunity. It could, for example, make Section 230 immunity contingent on whether a company has made reasonable efforts to identify and remove falsified, harmful content either at the upload stage or upon receiving notification about it after it is posted. 

However, such a legislative effort would certainly be met with stiff resistance from companies, as well as those who question whether such screening can be performed without an imposition of political or ideological bias.

Deep fakes do not always require a mass audience to achieve a harmful effect. From a national security and international relations perspective, the most harmful deep fakes might not flow through social media channels. Instead, they could be delivered to target audiences as part of a strategy of reputational sabotage. 

This approach will be particularly appealing for foreign intelligence services hoping to influence decision-making by people without access to cutting-edge detection technology.

Recommendations

The challenges of mitigating the threat of deep fakes are real, but that does not mean the situation is hopeless.

Enhancing current efforts by the National Science Foundation, Defense Advanced Research Projects Agency (DARPA), and Intelligence Advanced Research Projects Agency (IARPA) could spur breakthroughs that lead to scalable and robust detection capacities and digital provenance solutions. 

In the meantime, the current wave of interest in improving the extent to which social media companies seek to prevent or remove fraudulent content has pushed companies to take advantage of available detection technologies, flagging suspect content for further scrutiny, providing clear warnings to users, removing known deep fakes, and sharing such content in an effort to help prevent it from being reposted elsewhere, following a model used to limit the spread of child pornography. 

While by no means a complete solution, all of this would be a useful step 
forward.

The United States should also improve its efforts to combat hostile information operations that target US democracy and social cohesion, whether they feature deep fakes or not. One of the most potent tools available to the US government is its ability to issue targeted economic sanctions. This capacity has been used to a limited extent in response to Russian election interference in 2016. 

The executive branch needs to make clear that it can and will take similar measures anytime a foreign power attempts to distort US electoral processes, and that the reaction will be especially robust if the interference involves fraudulent materials along the lines of a deep fake. 

If the provocation is sufficiently serious, the US government could use cyber means to disrupt a hostile foreign information operation of this kind. In addition to traditional covert action, a series of provisions in the newly enacted John S. McCain National Defense Authorisation Act clarifies that US Cyber Command has the authority to use offensive cyber operations in response to such scenarios.

For some organisations and individuals, the best defense against deep fakes would be to establish highly credible alibis regarding where they have been and what they have been doing or saying. In practical terms, politicians and others with reputations to protect could have an increased interest in life-logging services. 

Such services would help insulate individuals and organizations from sabotage by ensuring they can prove where they were and what they were saying or doing at any given time. Service providers could sell life-logging equipment (such as tiny cameras) and authenticated storage services, similar to body cameras for police officers, and integrate these services with the screening mechanisms employed by major social media platforms, enabling rapid alibi-checking. 

However, this would increase the amount of surveillance in society and further erode notions of privacy, forcing some to choose whether their reputation and security is worth the price of privacy.

Finally, there is the simple option of spreading public awareness of the idea of deep fakes and encouraging skepticism of video and audio “evidence.” 

Yet this approach entails its own risk, the more people doubt their eyes and ears when faced with apparent evidence, the easier it becomes for liars to dispute legitimate proof of their misdeeds, a phenomenon known as the liar’s dividend. 

Those who hope to avoid accountability for legitimate video and audio evidence can exploit that skepticism. The cry of “fake news” will become the shout of “deep-fake news.” 

Problems such as declining receptivity to the idea of objective truth, or truth decay, and a growing tendency toward rejecting unwelcome evidence already exist, and a campaign to raise awareness about the danger of deep fakes would only pour fuel on that fire.

Notwithstanding these challenges, the public should be made aware that deep fakes exist.

Deep fakes are a profoundly serious problem for democratic governments and the world order. 

The United States should begin taking steps, starting with raising awareness of the problem in technical, governmental, and public circles so that policymakers, the tech industry, academics, and individuals become aware of the destruction, manipulation, and exploitation that deep fake creators could inflict.

Defense One:

You Might Also Read:

A Breakthrough In Video Analytics

AI Applied To Video Analytics

« School Training On Child Abuse And Cyber Security
Election Hacking Threatens US Mid-Terms »

CyberSecurity Jobsite
Perimeter 81

Directory of Suppliers

Cyber Security Supplier Directory

Cyber Security Supplier Directory

Our Supplier Directory lists 6,000+ specialist cyber security service providers in 128 countries worldwide. IS YOUR ORGANISATION LISTED?

North Infosec Testing (North IT)

North Infosec Testing (North IT)

North IT (North Infosec Testing) are an award-winning provider of web, software, and application penetration testing.

MIRACL

MIRACL

MIRACL provides the world’s only single step Multi-Factor Authentication (MFA) which can replace passwords on 100% of mobiles, desktops or even Smart TVs.

Resecurity, Inc.

Resecurity, Inc.

Resecurity is a cybersecurity company that delivers a unified platform for endpoint protection, risk management, and cyber threat intelligence.

Syxsense

Syxsense

Syxsense brings together endpoint management and security for greater efficiency and collaboration between IT management and security teams.

Certes

Certes

Certes is a pioneer in delivering cutting-edge security technology solutions, with a specific focus on Data Protection Risk Mitigation (DPRM).

INCIBE-CERT

INCIBE-CERT

INCIBE-CERT is the reference security incident response center for citizens and private law entities in Spain

Raz-Lee Security

Raz-Lee Security

Raz-Lee Security is the leading security solution provider for IBM Power i, otherwise known as iSeries or AS/400 servers.

Computer Forensic Services

Computer Forensic Services

Computer Forensic Services are digital evidence specialists. Practice areas include Information Security, e-Discovery, Law Enforcement Support and Litigation.

Woz U

Woz U

Woz U provides best-in-class technology training for Learners, Higher-Ed and Corporations. We focus on the most in-demand occupations such as Software Development, Data Science and Cyber Security.

SaferVPN

SaferVPN

SaferVPN is a Virtual Private Network services provider offering secure encrypted access to the internet.

X-Ways Software Technology

X-Ways Software Technology

X-Ways provide software for computer forensics, electronic discovery, data recovery, low-level data processing, and IT security.

Measured Insurance

Measured Insurance

Measured Insurance are bridging the gap between technology and Insurance using AI-Powered analytics that track clients’ exposure in real time to create smarter insurance products.

WhiteHawk

WhiteHawk

WhiteHawk is the first online Cyber Security Exchange. We help you understand your cyber risk and match you to tailored and affordable solutions.

ARIA Cybersecurity Solutions

ARIA Cybersecurity Solutions

The ARIA ADR Automatic Detection & Response solution was designed to find, verify, and stop all types of attacks - automatically and in real time.

F1 Security

F1 Security

F1 Security provides a family of web security solutions including web application firewalls, web shell detection solutions, and web shell scanners.

Input Output (IOHK)

Input Output (IOHK)

IOHK is one of the world's pre-eminent blockchain infrastructure research and engineering companies.

Sealing Technologies (SealingTech)

Sealing Technologies (SealingTech)

SealingTech is a leader in cutting edge research, products, engineering, and integration services in the Internet of Things, Edge, Machine Learning, Artificial Intelligence, and Cloud.

MicroAge

MicroAge

Powered by five decades of experience, lasting partnerships, client relationships, and the values that guide us daily, MicroAge is here to help you secure, accelerate, and transform your business.

Telindus

Telindus

Telindus is the strategic IT partner for the flexible organization of the future. We build optimal IT infrastructure with four components: networking, cloud, cybersecurity and data & AI.

Technology Innovation Institute (TII)

Technology Innovation Institute (TII)

TII is a UAE-based research center that aims to lead global advances in AI, robotics, quantum computing, cryptography and secure communications and more.