Major social media platforms are facing serious criticism for their inability to effectively manage harmful content pertaining to suicide and self-harm. A new report from the Molly Rose Foundation reveals shocking statistics: over 95% of such posts have gone unflagged on platforms like Facebook and Instagram.
The Molly Rose Foundation was established following the tragic death of 14-year-old Molly Rose, who took her own life after encountering harmful material online. This recent study scrutinized more than 12 million content moderation decisions across six prominent social media platforms.
According to findings, Pinterest and TikTok accounted for the majority of successful detections of suicide and self-harm posts, with each managing to effectively remove such content. Conversely, Meta's Instagram and Facebook barely registered, both managing just 1% of flagged content, with X (formerly Twitter) contributing only 0.14%.
Ian Russell, Molly's father and chairman of the foundation, voiced his disappointment, saying, "It's shocking to see major tech companies continue to sit on their hands and choose inaction over saving young lives." He strongly advocates for the government to advance the Online Safety Act to introduce stricter regulations.
The report highlights significant inconsistencies across platforms, particularly concerning their moderation policies. For example, Instagram's video feature Reels, which takes up half the app's usage time, yielded only one flagged post out of every 50 suicide-related submissions.
These results pinpoint not only failures to detect harmful content but also weak enforcement of existing rules. TikTok, for example, recognized nearly three million instances of self-harm posts, yet only two accounts were suspended.
Despite pledging action to restrict harmful content earlier this year, Meta has yet to fulfill its commitments, causing public outcry from concerned parents and advocates. Ian Russell expressed frustration, stating, "Parents will be rightly appalled at the negligence from social media giants. No ifs, no buts, assertive action is required."
The Molly Rose Foundation urges greater transparency and accountability from these platforms, particularly as regulations from the EU's Digital Services Act demand public disclosure of moderation practices. The foundation believes existing measures are woefully inadequate to protect users, especially children.
Research conducted previously highlighted how prevalent harmful content is on platforms like TikTok and Instagram; nearly half of the most engaged posts utilizing explicit suicide hashtags were flagged as unsuitable for minors. This alarming statistic has reignited calls for stronger legislation and intervention from authorities.
With more than four out of five UK adults and 84% of parents supporting tighter online safety legislation, the demand for meaningful action is clear. There is substantial pressure for the government to take decisive steps within the next two years of parliament.
Social media giants have countered concerns, asserting their policies prohibit any content glorifying self-harm or suicide. A spokesperson for Meta emphasized, "Content encouraging suicide and self-injury breaks our rules," indicating they are aware but struggling to enforce these regulations effectively.
Insufficient content moderation not only puts vulnerable individuals at risk but fosters environments where harmful material can thrive unchecked. The lack of decisive action from these platforms showcases their inconsistencies and prioritization of user engagement over safety.
Organizations, researchers, and parents alike stress the ultimate necessity for continuous improvement and proactive measures to shield users, particularly young ones, from dangerous content thriving online. Advocacy for the Online Safety Bill intensifies, citing firm legislative shifts as the only viable path to bridging the glaring gaps present today.
The Molly Rose Foundation’s continued efforts highlight the pressing need for tech firms to recognize and act on their social responsibility. With growing public support for safety regulations, change is increasingly becoming a pressing inevitability.