Course:JRNL503B/Deepfakes

From UBC Wiki

What are deepfakes?

A deepfake is a fabricated video, photo, or other piece of media made to look authentic with the use of artificial intelligence. The most popular use of deepfakes relates to videos. One of the more famous examples is this video of what looks like Barack Obama insulting Donald Trump, when in reality it’s filmmaker Jordan Peele’s voice synthesized to Obama’s face.

Deepfakes have been around for a while. The name comes from the portmanteau of Deep Learning and Fake, first used on a Reddit forum in 2017.

Visual effects studios have been superimposing actor’s faces onto stunt doubles for some time. Effects artists would go frame by frame with expensive video editing tools to painstakingly adjust lighting, blurs, unnatural shadows, and a number of other things. Now, the software is being used with machine learning. An algorithm takes each frame of a video of one person’s face (the target) and matches it to the other person’s face (the source), producing an output video of the target speaking the words of the source, or the source’s face superimposed on the target’s body.

“I refer to [a] deepfake as when a piece of synthetic media is used as a piece of mis[information] or disinformation.” - Nina Schick[1]

The source is filmed speaking and the algorithm analyzes the similarities and differences between the two videos, merging them together to create a believable video of the target image. Facial gestures are matched and ‘overlapped.’ The technology used to make them is surprisingly accessible, giving a whole new meaning to ’putting words into someone else’s mouth.’

This technology has already been weaponized as a tool of disinformation and blackmail. In August 2020, the U.S. Congressional Information Service published a report outlining the potential for deepfakes to be used by nefarious actors to “erode public trust, negatively affect public discourse, and even sway an election.”

The software, including programs like DeepFakeLab and Faceswap are free and open source. Anyone can download and run the software, and with cloud computing technology, anyone can rent the processing power required to render video images–a process that takes days and requires a lot of computational power.

Deepfakes of today are made possible with generative adversarial networks (GANs). In short, GANs are duelling computer programs that create new data from the same source data. This technology was invented by Ian Goodfellow and his team in 2014[2], and it positions each program within the dichotomy of a counterfeiter trying to fool police and a police officer trying to spot a counterfeit, thus forcing each program to use the same ‘learning’ technology our brains use when comparing two different but similar images and trying to distinguish the authentic one from the copy. The output is an extremely believable synthesized fake image.

Accessibility to these tools in combination with a growing repository of publicly available data has naturally led to an upsurge of highly-believable fabricated content online. The number of deepfakes videos online has grown exponentially and continues to increase.

External tools for spotting deepfakes

How to tell the difference between a deepfake and a real image

Deepfake technology isn’t perfect, and there are ways to spot a doctored video or a fake image. Still images are considerably harder to spot than videos, since a photo is only one edited frame, instead of several.

Facebook's AI division launched a deepfake detection challenge. The challenge was debuted at the NeurIPS 2019 conference, and was intended to help researchers innovate tools to detect deepfakes.

Signs of a deepfake

  • Fabric of clothing around the figure’s neck and face is different on the left versus right side of the image
  • Small glitches on the face including misplaced facial hair, freckles, flyaway hair
  • Lack of believable shadows between the figure in the foreground and the background
  • Unnatural hair movements

External tools for spotting deepfakes

  • Google reverse image search - allows you to see other copies of the image if they exist in a different context
  • YouTube Data Viewer or inVid - tells you the time/date of the uploaded content
  • Google Maps Street View, Google Earth - these can help corroborate locations by comparing background images
  • EXIFdata and Exiftool - these tools scrape the metadata from the video/image. By downloading the content and running it through either program you can get an idea of who posted it, where it comes from, and determine if it's real (though metadata can be scrubbed/altered)

What are deepfakes used for?

Despite the alarm, the majority of deepfakes have so far been created for entertainment with no intent of sowing geopolitical discord.

There have not been any examples of a deepfake being powerful enough to convince a newsroom that they’re real. Although plenty have gone viral and are shared widely on social media, none have reached the heights of influencing governments. Yet.

Entertainment

The majority of deepfakes that you find on YouTube and Reddit are funny face-swaps of celebrities. Most are marked with a [DeepFake] tag and are seemingly harmless forms of comedy, often involving the actors Nicolas Cage and Rowan Atkinson.

Some of the most infamous examples include:

  • Dr. Phil on Dr. Phil


  • Sylvester Stalone in Home Alone


  • Mr. Bean as Charlize Theron


  • Ron Swanson as Wednesday Adams


  • Nicolas Cage as just about everyone


Politics

Misrepresenting what someone says or does is nothing new. But deepfake technology makes the potential for harm higher than ever. There are many deepfake videos involving Donald Trump, but none have yet fooled any other heads of state.

U.S. Senator Ben Sasse said that deepfakes were “likely to send American politics into a tailspin” and introduced a bill that would make it a crime to create or distribute deepfakes with malicious intent.

Despite the panic, there is currently no evidence that deepfakes were used in the 2020 U.S. elections to influence either campaign.

There have been several doctored videos that have made headlines, including one of Nancy Pelosi appearing intoxicated, and journalist Jim Acosta, which was changed to make it seem like he ‘chopped’ the arm of an intern who was taking a mic from him.

One of the more nefarious ways deepfake technology is being used to influence journalism is with AI bots creating fake Twitter profiles of reporters.  

Business

Some companies are in the business of selling fake people.

The New York Times reports that, “on the website Generated.Photos, you can buy a ‘unique, worry-free’ fake person for $2.99. If you just need a couple of fake people — for characters in a video game, or to make your company website appear more diverse — you can get their photos for free on ThisPersonDoesNotExist.com.”

Deepfakes and pornography

“Deepfakes are rarely used to help usher in a dystopian political nightmare where fact and fiction are interchangeable: They exist to degrade women.” - Aja Romano

In 2019, Sensity.ai released a report which found that 96% of deepfake videos are simulating porn of female celebrities without their consent[3]. Most appeared on the Reddit thread r/deepfake which has now been banned for violating Reddit’s policy against sharing content that contained involuntary pornography.

“This is all part and parcel of the broader abuse and harassment that women have to deal with in the online environment,” says Nina Jankowicz, author of How to Lose the Information War.

It’s not just famous women anymore. An AI bot can use a single image of any woman to create pornographic images. Earlier this year, an AI bot was discovered trawling the messaging app Telegram for women’s user images and turning them into porn.

It has opened the floodgates of deepfaked revenge porn.

Noelle Martin found herself a victim of deepfake porn. She was emailed graphic videos of a porn star with her face. “It was convincing, even to me," she writes. The ordeal led to her starting a campaign for the Australian government to do more to tackle the issue of involuntary pornography.

Rana Ayyub is an investigative journalist and sexual violence activist, who also became the victim of deepfake pornography bearing her image. The abuse, intimidation and sexual blackmail that deepfakes can be used for results in self-censorship and builds an atmosphere of fear around female public figures.

“Perhaps fake porn is not critically addressed in the news media because it does not dupe or disempower the editors and executives who set media agendas, which disproportionately focus on men. While political deep fakes threaten their individual truth-seeking power, fake porn appears to reinforce it.” - Sophie Maddocks[4]

The greatest threat that deepfakes pose to journalism is in their ability to threaten and potentially silence female journalists.

Threats to truth and transparency

Experts like Ashish Jaiman, Director of Technology and Operations at Microsoft, call deepfakes an “imminent threat”[5] to society. There are also concerns around deepfakes threatening national security as well as creating significant economic losses[6]. The harm caused by deepfakes can be broadly categorized into:

  • Disinformation (and related impact on fake news, national security, economic conduct, social distrust and the liar’s dividend[7])
  • Weaponized harm (to blackmail, bully or create emotional duress for individuals, especially using deepfake porn)
  • Threats to cybersecurity (by creating models that promote human error.)
  • Intellectual property abuse (by deepfaking artists and releasing new art by "them.")

Easier access to cloud computing and the democratization of AI allows individuals with minimal resources to create deepfakes. Improvements[8] in Generative Adversarial Networks not only make for more realistic deepfakes but also outpace efforts to spot and regulate them.  

Major software techniques to identifying deepfakes

There are two types of solutions[9] to detecting deepfakes.

  • Detection

Detection software and tools are created to be able to spot whether a piece of media is a deepfake after its creation. Detection tools check the media for audio inconsistencies, inaccurate shadows, variation in physical features and other such differences that the human eye can’t see.

Due to the nature of GANs, the creation of detection tools requires large datasets and computing power and runs the risk of quickly becoming obsolete.

Major technology companies like Facebook, Twitter, Google, Microsoft, etc) and government organisations (like the U.S. Defense Advanced Research Projects Agency) are the few places where the funding and creation of detection technology can be sustained in the long run.

  • Authentication and provenance

Authentication and provenance tools are considered a long-term solution to combat harmful deepfakes.

Using metadata, author tags and an attempt to create a single certification database[10] allows platforms to spot a deepfake before its published. These solutions rely on proofing the source and reliability of all published media, making a deepfake stand out.

Since these solutions do not require creation of computation intensive software, they are less expensive and not as resource heavy. However, only a wide adoption of a standardized framework by media creators, generators and publishers will be an effective solution against deepfakes.      

Some apps are also building databases of verified media which can be used to source and detect deepfakes.

What can journalists do about deepfakes?

Deepfakes pose a dual threat to journalists and journalism as a whole. Their capacity for disinformation, along with their technical prowess, means journalists will have to develop new tools and processes[11] to verify videos which could be deepfakes, or deepfakes that become newsworthy themselves.

They will also need to investigate claims where true media is accused of being a deepfake.

Meanwhile, the threat of deepfakes could lead to an erosion of trust as a whole, where potentially, any piece of media could be a deepfake, increasing polarization and reducing trust in the news media.    

"We have to avoid playing into the hands of people who want to call everything ‘fake news’ and to technology solutions that will completely substitute a technical signal for human judgement, rather than complement human judgement. Yet we do have to prepare." - Sam Gregory, Program Director, WITNESS.org

While there have been no major incidents of inaccurate reporting due to a deepfake, or prolonged investigations into the authenticity of media suspected to be a deepfake, there is a need for newsrooms to build a collaborative approach to dealing with them.

Major newsrooms have signed up with technology companies to combat disinformation, specifically, deepfakes. The New York Times, CBC and the BBC have partnered with Microsoft’s Project Origin to test the company’s authenticity technology.

The same companies are also reportedly testing the use of deepfakes to protect journalists reporting in hostile environments. In a positive twist, a deepfake could be employed to change the journalist’s identity while allowing them to report through audio or video.

The Wall Street Journal has formed a committee of 21 newsroom members across different departments that will help identify disinformation, including “AI-synthesized media” (deepfakes.)    

Coverage of deepfakes

A collage of different headlines from major media outlets of their coverage of deepfakes. Clockwise from top left: The Guardian (Ian Sample,) The Rolling Stone (EJ Dickson,) The New York Times (Regina Rini,) The Wall Street Journal (Hilke Schellmann,) and The Washington Post (Drew Harwell).

The mainstream media has been responsive to the threat of deepfakes with coverage appearing in The New York Times, The Wall Street Journal, The Washington Post, The Guardian, the BBC, and the CBC.

Most outlets focused on, yet struggled to specify, how much of a problem deepfakes can be. While accurately pointing out the potential harm, they were less certain on where its next realistic target, and consequential damage, would be. More recent coverage discusses whether the deepfake threat to the U.S. election was overblown by the media.       

The media has also done a reasonable job of covering the harm done by deepfakes against women, usually in response to inciting incidents.

While coverage focuses on the incident and the enormity of harm, it usually lacks in providing context to solutions or proposed legal regulation in this area, like the DEEPFAKES Accountability Act (2019). The Act proposes many regulatory changes, one of which is requiring all deepkaes to be identified as such by their creators.

More recent coverage from smaller, technology focused outlets, have criticized the DEEPFAKES Act for the many loopholes it fails to address. One of the key criticisms is that its requirements are naive and targeting the wrong demographic of deepfake creators. It correctly points out that "if a creator of a piece of media is willing to put their name to it [a deepfake] and document that it is fake, those are almost certainly not the creators or the media we need to worry about." However it also mentions how hard creating effective legislation would be because "[technology] frequently is some distance ahead of the law, not just in spirit but in letter."

Other niche websiteshave also gone on to put coverage from legacy outlets into the context of proposed policy moves in the U.S. and around the world.

References

  1. Schick, Nina, guest (17 October 2020). "Making Sense with Sam Harris Ep. 240 The Information Apocolypse". Sam Harris.
  2. Goodfellow, Ian; David Warde-Farley, Jean Pouget-Abadie; Mehdi Mirza; Bing Xu; Sherjil Ozair; Aaron Courville; Yoshua Bengio. "Generative Adversarial Nets" (PDF). Universite de Montreal: 1–9.
  3. The State of Deepfakes: Landscape, Threats, and Impact, Henry Ajder, Giorgio Patrini, Francesco Cavalli, and Laurence Cullen, September 2019.
  4. Maddocks, Sophie (June 2020). "'A Deepfake Porn Plot Intended to Silence Me': exploring continuities between pornographic and 'political' deep fakes". Porn Studies.
  5. Jaiman, Ashish. Director of Technology and Operations, Microsoft. Personal interview. November 11th, 2020
  6. Fraga-Lamas, Paula and Fernandez-Caramés Tiago M. (2019) ‘Fake News, Disinformation, and Deepfakes: Leveraging Distributed Ledger Technologies and Blockchain to Combat Digital Deception and Counterfeit Reality’. Faculty of Computer Science, Universidade da Coruna.
  7. Chesney, Robert and Citron, Danielle Keats, (July 14, 2018). 'Deep Fakes: A Looming Challenge for Privacy, Democracy, and National Security.' 107 California Law Review 1753 (2019), U of Texas Law, Public Law Research Paper No. 692, U of Maryland Legal Studies Research Paper No. 2018-21.
  8. Yuezun Li, Ming-Ching Chang and Siwei Lyu (2018) ‘In Ictu Oculi: Exposing AI Generated Fake Face Videos by Detecting Eye Blinking’. Computer Science Department, University at Albany, SUNY.
  9. Jaiman, Ashish. Director of Technology and Operations, Microsoft. Personal interview. November 11th, 2020
  10. England, Paul et. al, (2020) ‘AMP: Authentication of Media via Provenance’. Microsoft.
  11. Sohrawardi, Saniat et. al., (2020) 'DeFaking Deepfakes: Understanding Journalists' Needs for Deepfake Detection'. USENIX.