Meta fails to moderate self-harm images on Instagram
- Danish researchers established a private self-harm network on Instagram to test moderation.
- Despite claims of removing harmful content, none of the self-harm images shared during the study were taken down.
- The study raises concerns about Meta's commitment to user safety, particularly among vulnerable teenagers.
In a recent study conducted in Denmark, researchers investigated the presence of self-harm content on Instagram and found severe flaws in the platform's content moderation. The researchers created a private network focused on self-harm, including fake profiles of individuals as young as 13 years old, and shared graphic content that escalated in severity over a month. The results were alarming; not a single image was removed by Instagram's moderation team during the entire investigation. This led to assertions that Meta does not effectively manage small private groups, which could be concealing harmful activities from parents and authorities. According to Digitalt Ansvar, the organization behind the research, while Meta claims to remove around 99% of harmful content proactively, their study suggested a significant disconnect between this claim and the reality experienced within the platform. The organization developed its own AI tool to assess the content, finding that it could automatically detect 38% of self-harm images and 88% of the most severe types. The lack of action from Meta contradicts their public assurances about tackling self-injury content. Psychologist Lotte Rubæk, an ex-member of Meta’s global suicide prevention expert group, echoed concerns over the findings, noting that the zero removal rate of the 85 posts shared during the study was shocking. She highlighted how failing to address such explicit images can potentially trigger vulnerable youths, especially young women, thus contributing to the escalating rates of self-harm and suicides among this demographic. The urgency of the findings reflects a broader issue concerning the safety and mental well-being of teenagers using social media platforms. This research raises critical questions about Meta’s adherence to regulations like the EU’s Digital Services Act, which mandates large digital platforms to identify systemic risks that could negatively impact mental health. The failure to effectively moderate content on platforms like Instagram has significant implications, prompting discussions about the responsibilities of social media companies in protecting their users, especially minors, from harmful content. There remains significant public and regulatory pressure on Meta to ensure that algorithms and moderation practices accurately protect vulnerable populations from the proliferation of self-harm content and its deeply detrimental effects.