Skip to main content



commentary Commentary

Commentary: We need to talk about the mental health of content moderators

Content moderators spend hours sifting through distressing content which repeated exposure can have serious consequences including PTSD, says one University of Melbourne observer.

Commentary: We need to talk about the mental health of content moderators

Facebook moderators sift through hundreds of examples of distressing content during each eight hour shift. (Photo: AFP/Christophe SIMON)

MELBOURNE: Selena Scola worked as a public content contractor, or content moderator, for Facebook in its Silicon Valley offices. She left the company in March after less than a year.

In documents filed last week in California, Scola alleges unsafe work practices led her to develop post-traumatic stress disorder (PTSD) from witnessing “thousands of acts of extreme and graphic violence”.

Facebook acknowledged the work of moderation is not easy in a blog post published in July. In the same post, Facebook’s Vice President of Operations Ellen Silver outlined some of the ways the company supports their moderators:

All content reviewers — whether full-time employees, contractors, or those employed by partner companies — have access to mental health resources, including trained professionals onsite for both individual and group counselling.

But Scola claims Facebook fails to practice what it preaches. Previous reports about its workplace conditions also suggest the support they provide to moderators isn’t enough.

Scola’s legal action is not the first of its kind. Microsoft have been involved in a similar case since December 2016 involving two employees who worked in their child safety team.

In both cases the plaintiffs allege their employer failed to provide sufficient support, despite knowing the psychological dangers of the work.

Both Microsoft and Facebook dispute the claims.


Facebook moderators sift through hundreds of examples of distressing content during each 8-hour shift.

They assess posts including, but not limited to, depictions of violent death – including suicide and murder – self-harm, assault, violence against animals, hate speech and sexualised violence.

An employee walks past a Facebook logo at its new London headquarters (Photo: AFP/Daniel LEAL-OLIVAS)

READ: Facebook’s fake news crisis is a product of Silicon Valley culture, a commentary

Studies in areas such as child protection, journalism and law enforcement show repeated exposure to these types of content has serious consequences. That includes the development of PTSD. Workers also experience higher rates of burnout, relationship breakdown and, in some instances, suicide.


Industries including journalism, law and policing have invested a significant amount of thought and money into best practice policies designed to protect workers.

Those working in child safety opt-in to the work rather than cases being assigned in Australia, for instance. They are then required to undertake rigorous psychological testing to assess if they are able to emotionally compartmentalise the work effectively. 

Once working, they have regular mandated counselling sessions and are routinely reassigned into other areas of investigation to limit the amount of exposure.

The tech industry has similar guidelines. In fact, Facebook helped create the Technology Coalition, which aims to eradicate online child sexual exploitation. 

In 2015, the coalition released its Employee Resilience Guidebook, which outlines occupational health and safety measures for workers routinely viewing distressing materials. While these guidelines are couched as specific to workers viewing child pornography, they are also applicable to all types of distressing imagery.

The guidelines include “providing mandatory group and individual counselling sessions” with a trauma specialist, and “permitting moderators to opt-out” of viewing child pornography.

(Photo: Unsplash/Alexandre Godreau)

The guidelines also recommend limiting exposure to disturbing materials to four hours, encouraging workers to switch to other projects to get relief, and allowing workers time off to recover from trauma.


Having support available doesn’t necessarily mean staff feel like they can actually access it. Most of Facebook’s moderators, including Scola, work in precarious employment conditions as outside contractors employed through third-party companies.

Working under these conditions has been shown to have a detrimental impact on employee well-being. That’s because these kinds of employees are not only less likely to be able to access support mechanisms, they often feel doing so will risk them losing their job. 

In addition, low pay can lead to employees being unable to take time off to recover from trauma.

Insecure work can also impact one’s sense of control. 

READ: Tip of the iceberg, when ‘disguised PMEs’ don’t get overtime pay and fair treatment, a commentary

Moderators have little to no control over their work flow. They do not control the type of content that pops up on their screen. They have limited time to make decisions, often with little or no context. And, they have no personal say in how those decisions are made.

According to both the filing, and media reports around Facebook’s moderator employment conditions, employees are under immense pressure from the company to get through thousands of posts per day. They are also regularly audited, which adds to the stress.


Adequate workplace support is essential for moderators. Some sections of the industry provide us with best case examples. In particular, the support provided to those who work in online mental health communities, such as Beyond Blue in Australia, is exemplary and provides a good blueprint.

We also need to address the ongoing issue of precarity in an industry that asks people to put their mental health at risk on a daily basis. 

READ: We still fail to understand that some people are more prone to mental illness, a commentary

(File photo: AP) Depression can present itself as a form of extreme sadness, with a risk of self-destruction. (Photo: AP)

This requires good industry governance and representation. For instance, Australian Community Managers have recently partnered with Australia's Media, Entertainment and Arts Alliance to push for better conditions for everyone in the industry, including moderators.

As for Facebook, Scola’s suit is a class action. If it’s successful, Facebook could find itself compensating hundreds of moderators employed in California over the past three years. 

It could also set an industry-wide precedent, opening the door to complaints from thousands of moderators employed across a range of tech and media industries.

Jennifer Beckett is lecturer in media and communications at the University of Melbourne. This commentary first appeared on The Conversation.

Source: CNA/nr


Also worth reading