Skip to content →

Courtney Jones Posts

It’s Time to Stop Calling It “Deepfake Porn”

Lately, the news has been full of stories of women being subjected to a new form of image-based sexual abuse (or IBSA, for short), perpetrated using artificial intelligence technologies. Somewhat recently, fake videos of Taylor Swift went viral on X which depicted her performing sexual acts, before being taken down and her name becoming temporarily unsearchable on the platform. A quick Google search of “deepfake pornography” will yield countless results of women sharing their stories of being victimised using this technology and advocates calling on lawmakers to do something about it.

But here’s the thing. The language that we use to talk about this phenomenon is a problem.

(c)
(c) Igor Omaliev via Unsplash

When there was an explosion in the use of mobile phones by teens and young adults, and social media, we saw a rise in conversations discussing “revenge porn”. However, advocates rightly highlighted that this language was harmful for two reasons. First, because calling it revenge pornography insinuates that the victim did something to deserve it – it implies it was her fault. Second, referring to this form of abuse as “pornography” is not right either. Some (though many feminists would disagree) would consider pornography to be a form of entertainment, and the non-consensual sharing of nude or sexual images certainly is not that.

Similarly, in a LinkedIn post during August 2024, CEO of the National Center on Sexual Exploitation in the United States, Dawn Hawkins, highlighted problems with using the term intimate image abuse or non-consensual intimate imagery. She writes:

“Referring to IBSA as “intimate” imagery is a gross mischaracterization. The word “intimate” implies consent and privacy, which are wholly absent in cases of IBSA. This term fails to capture the exploitation, extortion, harassment, and humiliation that victims endure. Let’s be clear: IBSA is a form of sexual assault and public humiliation, not an intimate act.”

So why, then, do we use the term “pornography” when referring to this form of deepfake imagery?

I argue that we should not be referring to this form of abuse as “deepfake pornography” when there are so many other options. The non-consensual creation and/or sharing of nude or sexual imagery is image-based sexual abuse (abbreviated IBSA), and we should be using language to describe it this way.

Victoria Rousay has proposed referring to these images as sexual deepfakes, allowing us to discuss the sexual nature of the image without referring to it as “pornography”. This is a great option able to encompass both non-consensual imagery as well as that made with the consent of the person/people depicted.

However, I would like to propose two alternatives. The first proposed alternative is non-consensual synthetic sexual imagery (abbreviated NCSSI). Despite its wordiness, I think this is one option that allows us to appropriately encapsulate what has been done – synthetic (fake) sexual imagery was created (or shared) without the consent of those depicted. A second alternative, which may be preferable for those of us who love abbreviations, would be AI-facilitated or AI-generated image-based sexual abuse (AI-IBSA).

Rebekah Wells, who founded Women Against Revenge Porn, said when discussing image-based sexual abuse:

“It has been called “revenge porn,” “involuntary pornography” and “nonconsensual pornography.” But using these terms is like calling rape “involuntary sex.” It simply doesn’t reflect the emotional, psychological and physical costs. Revenge porn is cyberrape, and we should call it as such.”

Wells suggests using the term cyberrape to refer to seemingly all forms of image-based sexual abuse, and this would extend beyond referring to only images generated or altered using AI or deepfake technologies. However, what is clear is this: we need to stop referring to non-consensual deepfakes as “pornography”.

Leave a Comment

What’s Missing From the Conversation on AI? Women.

Unless you’ve been living an entirely disconnected life for the past few years, you will have heard of large language model applications like ChatGPT. There is no doubt that artificial intelligence (AI) is rapidly becoming increasingly intelligent. However, there is something we have been leaving entirely out of the conversation – the impact of this technology on women.

(c) Gertrūda Valasevičiūtė

You may have already heard of deepfake technology – that which allows someone to create false videos of anyone, but mostly targets celebrities and politicians. This same technology can be used to create deepfake pornography, and make it appear as though a particular woman has participated in a pornographic film – even though she never has. There is another type of similar technology which aims to allow users toto create nude images of women through the use of AI. One such app called DeepNude was pulled from app stores after public outcry on social media. The difference between the two is that deepfake pornography superimposes another person’s face onto the body of a porn actor, whereas nudifying tools aim to generate “nudified” images from fully or partially clothed images uploaded by a user.

An MP in the United Kingdom, Maria Miller, had in 2021 called to ban these tools. In speaking with the BBC in August 2021, she said that a debate needed to occur on “whether nude and sexually explicit images generated digitally without consent should be outlawed”.

So-called “revenge porn” is already a criminal offence in Canada under section 162.1 of the Criminal Code, where an intimate image is defined as “a visual recording of a person made by any means including a photographic, film or video recording” where this person is depicted exposing their sexual organs or engaging in sexual activity.

Although this specifies that these images could be made by any means, it is yet to be determined whether this law would apply to images generated with the use of artificial intelligence, as this has yet to be tested by Canadian courts.

It is difficult to find research on deepfake pornography or nudifying tools in Canada, but a report coming from the U.K.’s Law Commission pointed to a study that found that 100% of the victims of deepfake pornography are women. In fact, “nudifying tools” will generate female genital organs on any person uploaded, regardless of their sex. The technology assumes that the person you are attempting to nudify is a woman.

Technologies like those used to create deepfake pornography are already impacting the lives of women. Just this February, a streaming star on Twitch was victimised by deepfake porn technology, and noted that her fans thought the video is real. What Sweet Anita said in her interview with the New York Post is true – this could happen to anyone. Any woman wiho has uploaded images to a social media account, or has had images uploaded of her on her friend’s accounts, can be “deepfaked” or “nudified”. Since this particular incident, Twitch has updated its policy so that anyone “intentionally promoting, creating, or sharing” deepfake pornography, or what they refer to as “synthetic non-consensual exploitative images” could be “indefinitely suspended” upon a first offence.

Twitch’s move to indefinitely suspend those who share deepfake pornography, or other non-consensual sexually exploitative material on their website, is a step in the right direction. However, what Twitch is unable to do is address the root of the problem – the proliferation of the technology that allows such victimisation of women.

As we live in a world that is becoming increasingly digitised and as AI continues to proliferate, we, as a society, need to have conversations about how this technology can have a very real and negative impact on women’s lives, and our lawmakers need to act accordingly. We must put pressure not only on our governments but on technology companies to ensure the products they are creating and improving are not then weaponised against half of the global population.

Leave a Comment

Thoughts on International Women’s Day

Given that the theme for this year’s International Women’s Day is #DigitAll: Innovation and technology for gender equality, it only seems fitting that I’m spending my time attending sessions online at the 67th United Nations Commission on the Status of Women.

As I write this, I’m listening to a parallel event conversation circle on freedom of opinion and expression. Many of the discussions are centred on whether online social platforms allow women to freely express ourselves, or whether these platforms are providing safe online spaces for women. The overwhelming response is that no, they don’t.

I’ve heard from countless women today mentioning that the second they started making political points online, whether they were feminist, anti-racist, pro gay rights, etc., they started to receive online harassment and abuse. They also mentioned the inherently gendered nature of that abuse.

It’s very clear that technology is designed by men, for men, and that women are actively prevented from accessing online spaces and sharing our opinions freely. Given that an overwhelming amount of discussions, especially political ones, take place in online forums, this means women are actively being kept from political discussions and processes.

This International Women’s Day, it’s important that we think about how to change online platforms and online behaviour to create spaces that women can participate in safely. On an individual level, this looks like calling out online abuse and trolling when we see it. We especially need men to do this. On a broader, institutional level, we need our governments to regulate online spaces in a way that allows for women to be able to participate in them safely without hindering freedom of expression.

Access to these online spaces is particularly important so that we can be informed about how women’s rights are being impacted on a global scale. Much of the reporting of the protests in Iran surrounding the death of Mahsa Amini, for example, are taking place online. The Centre for International Governance Innovation has discussed how the women in Iran are taking to social media in fighting for their human rights.

If you’re a man reading this, consider how your online behaviour could change to create a safer space for online women, including calling out threats of violence and the online trolling of women. If you’re a woman reading this, do not allow them to silence you. Your voice matters, and the more voices we have fighting the good, feminist fight, the better.

Happy International Women’s Day.

Leave a Comment