The Times Australia
Google AI
The Times World News

.

Online abuse against women is rife, but some women suffer more – and we need to step up for them

  • Written by Hannah Klose, Teaching Associate/PhD Candidate, Monash University
Online abuse against women is rife, but some women suffer more – and we need to step up for them

Women online suffer a disproportionate amount of harm and abuse, but it isn’t all based on their gender. This “cyber violence” is also shaped by a range of other intersecting[1] factors[2] such as race, religion, class, caste and disability.

Our ongoing research involves collecting case studies from both India[3] and Australia to understand how various marginalised identities can impact young women’s experiences of online violence, and how social media companies – including Facebook, Twitter and Instagram – aren’t doing enough to stop it.

India is a rich case study for this research, as it’s a country where women have many different expressions of identity in large numbers – and where there remains a lot of racial, religious and social tension across society.

However, although Australia and India have significantly different cultures, women in both countries fall victim to online[4] crimes[5], including cyber stalking and cyber harassment. And those with marginalised identities have to deal with more more stigma and targeting.

What’s worse is platform content moderators are failing to recognise this cyber violence – often because they don’t understand the nuance and contexts in which[6] stigmas operate.

What is cyber violence?

Cyber violence[7] can be understood as harm and abuse facilitated by digital and technological means.

In 2019, there was a 63.5% increase[8] in the number of cyber violence cases being reported in India, compared to 2018. There has since been a further rise in cases against women from marginalised communities, including Muslim and Dalit women.

One prominent example is the “Bulli Bai[9]app[10], which turned up on GitHub in July last year. The app developers used the images of some 100 Muslim women[11] without their permission, to put them up “for sale” in a fake auction. The purpose was to denigrate and humiliate Muslim women in particular.

This is mirrored in Australia. Young Indigenous women are susceptible to being on the receiving end of cyber violence which not only targets them by gender, but also race.

A 2021 research report by eSafety[12] found Aboriginal and Torres Strait Islander women felt victimised by racist and threatening comments made online, usually in public Facebook groups. They also reported feeling unsafe and having their mental health significantly impacted.

Another example comes from New South Wales Greens Senator Mehreen Faruqi[13], who has received extremely[14] high[15] levels of online abuse as Australia’s first female Muslim senator. Speaking on behalf of women from marginalised backgrounds, Faruqi said[16]:

It is based on where I come from, what I look like, my religion.

Young women with marginalised identities

Research[17] on cyber violence against women in India reveals how hatred towards certain religions, races and sexual orientations can make gender-based violence even more harmful.

When women express their opinions or post pictures online, they are targeted based on their marginalised identities. For instance, Kiruba Munusamy, an advocate practising in the Supreme Court of India, received racist and caste-based slurs[18] for speaking out about sexual violence online.

And women with marginalised identities continue be victimised online[19], despite attempts to control this.

Read more: A better way to regulate online hate speech: require social media companies to bear a duty of care to users[20]

Take Australia’s “Safety by Design” framework[21], developed by the eSafety commissioner. Despite having some gathered traction in the past few years, it remains a “voluntary” code that encourages technology companies to prevent online harm through product design.

In India, hate speech against Muslims in particular has been on the rise[22]. India has laws (albeit flawed) that[23] can be used to deal with online abuse, but better implementation is needed.

With a Hindu majority, and radicalisation, it can be difficult to report incidents. Victims are concerned about safety and secondary victimisation, wherein they may face further abuse as a result of reporting a crime.

Read more: Why Modi's India has become a dangerous place for Muslims[24]

It’s hard to know the exact amount of cyber violence perpetrated against women with marginalised identities. Yet it’s clear these identities are linked to the amount of, and type of, abuse women face online[25].

One study by Amnesty International found[26] Indian Muslim women politicians faced 94.1% more ethnic or religious slurs than women politicians of other religions, and women from marginalised castes received 59% more caste-based slurs than women from more general castes.

We’ve long understood the need for an intersectional approach to feminism. We now need the same approach to protecting women’s safety online. Shutterstock

Recognition in platform design

Five years ago, Amnesty International submitted a report[27] to the United Nations highlighting the need for moderators to be trained in identifying gender-related and identity-related abuse on platforms.

Similarly, in 2019 Equality Labs in India[28] published an advocacy report discussing how Facebook failed to protect people from marginalised Indian communities[29]. This is despite Facebook having caste, religion and gender as “protected” categories under hate speech guidelines.

Yet in 2022 social media companies and moderators still need to do more to approach cyber violence[30] through an intersectional lens. While platforms have country-specific moderation teams, moderators will often lack cultural competency and literacy on matters of caste, religion, sexuality, disability and race. There could be various reasons for this, including a lack of diversity among staff and contractors[31].

In a 2020 report by Mint[32], one moderator working for Facebook India said she’s expected to achieve an accuracy report of 85% minimum to keep her job. In practise, this means she can’t spend more than 4.5 seconds on content being reviewed. Such structural issues can also contribute to the problem.

The way forward

In March 2022, the eSafety Commission in Australia[33] joined a global partnership[34] to end cyber violence against women. But a great deal of work still needs to be done.

Content moderation can be complex, and requires collective expertise from communities and advocates[35]. One way forward is to enforce transparency, accountability and resource allocation to build solutions within social media companies.

In November last year, the Australian government released the draft of a bill[36] aimed at holding social media companies accountable for content posted on their platforms, and protecting people from trolls.

It’s anticipated these regulations will ensure platforms are held responsible for harmful content that affects users.

Read more: Leigh Sales showed us the abuse women cop online. When are we going to stop tolerating misogyny?[37]

References

  1. ^ intersecting (www.amnesty.org)
  2. ^ factors (oxfamilibrary.openrepository.com)
  3. ^ India (journals.sagepub.com)
  4. ^ online (www.theguardian.com)
  5. ^ crimes (www.coe.int)
  6. ^ nuance and contexts in which (journals.sagepub.com)
  7. ^ Cyber violence (eige.europa.eu)
  8. ^ 63.5% increase (timesofindia.indiatimes.com)
  9. ^ Bulli Bai (timesofindia.indiatimes.com)
  10. ^ app (www.bbc.com)
  11. ^ 100 Muslim women (www.ibtimes.com.au)
  12. ^ research report by eSafety (www.esafety.gov.au)
  13. ^ Mehreen Faruqi (www.abc.net.au)
  14. ^ extremely (www.sbs.com.au)
  15. ^ high (www.theguardian.com)
  16. ^ said (digitallibrary.un.org)
  17. ^ Research (journals.sagepub.com)
  18. ^ racist and caste-based slurs (genderit.org)
  19. ^ online (link.springer.com)
  20. ^ A better way to regulate online hate speech: require social media companies to bear a duty of care to users (theconversation.com)
  21. ^ “Safety by Design” framework (www.esafety.gov.au)
  22. ^ on the rise (thediplomat.com)
  23. ^ laws (albeit flawed) that (www.aljazeera.com)
  24. ^ Why Modi's India has become a dangerous place for Muslims (theconversation.com)
  25. ^ women face online (www.europarl.europa.eu)
  26. ^ found (www.amnestyusa.org)
  27. ^ report (www.ohchr.org)
  28. ^ Equality Labs in India (static1.squarespace.com)
  29. ^ communities (www.hup.harvard.edu)
  30. ^ approach cyber violence (dl.acm.org)
  31. ^ lack of diversity among staff and contractors (www.equalitylabs.org)
  32. ^ Mint (www.livemint.com)
  33. ^ eSafety Commission in Australia (www.esafety.gov.au)
  34. ^ global partnership (www.state.gov)
  35. ^ communities and advocates (unidir.org)
  36. ^ the draft of a bill (www.ag.gov.au)
  37. ^ Leigh Sales showed us the abuse women cop online. When are we going to stop tolerating misogyny? (theconversation.com)

Read more https://theconversation.com/online-abuse-against-women-is-rife-but-some-women-suffer-more-and-we-need-to-step-up-for-them-183646

Times Magazine

Why Car Enthusiasts Are Turning to Container Shipping for Interstate Moves

Moving across the country requires careful planning and plenty of patience. The scale of domestic ...

What to know if you’re considering an EV

Soaring petrol prices are once again making many Australians think seriously[1] about switching ...

Epson launches ELPCS01 mobile projector cart

Designed for the EB-810E[1] projector and provides easy setup for portable displays in flexible ...

Governance Models for Headless CMS in Large Organizations

Where headless CMS is adopted by large enterprises, governance is the single most crucial factor d...

Narwal Freo Z10 Robotic Vacuum and Mop Cleaner

Narwal Freo Z10 Robotic Vacuum and Mop Cleaner  Rating: ★★★★☆ (4.4/5) Category: Premium Robot ...

Shark launches SteamSpot - the shortcut for everyday floor mess

Shark introduces the Shark SteamSpot Steam Mop, a lightweight steam mop designed to make everyda...

The Times Features

South Australian Nationals to open up local oil from Great Australian Bight

Amid out-of-control inflation and impacts from the Middle East conflict, The South Australian Na...

How does your super balance compare to other people your age?

If you have ever checked your super balance and wondered whether you are “behind” for your age, ...

Why Farrer is a key test for One Nation vs the Coalition

The Farrer by-election[1] on May 9 will be a major test for new Liberal leader Angus Taylor and ...

Leader of The Nationals Senator Matt Canavan Rockhampton press conference

Well thank you ladies and gentlemen. Thank you for coming out, this morning and thank you very muc...

Chester to elevate food security issue in Canberra

Elevating the issue of food and fibre security to a matter of national importance will be the prim...

Interior Design Ideas for Open Plan Living Spaces

Open plan living has become one of the most popular layout choices in modern homes. By removing wa...

Matt Canavan is keen on income splitting. Here’s what it would mean for couples

Newly elected Nationals leader Matt Canavan has proposed[1] allowing couples with dependent chil...

Custom Homes vs Project Homes: What’s the Difference?

When building a new home, one of the first and most important decisions you’ll make is whether to ...

Tech companies are blaming massive layoffs on AI. What’s really going on?

In the past few months, a wave of tech corporations have announced significant staff cuts and ...