Could New Tools Help Prevent Suicide and Self-Harm?
Facebook has announced a set of new tools to help detect suicidal intent messages, live videos, and posts. It has released these tools as a response to a growing international public health problem: one person commits suicide every 40 seconds, according the World Health Organization (WHO) Report Preventing Suicide: A Global Imperative.
There are three reasons why social media is in a unique position to detect, as well as to help prevent, suicide and other grave self-harm.
In addition, there exists a phenomenon of suicides on social media itself, whereby people have actually broadcast their own suicides to online audiences.
Incidences of mortal self-harm that are influenced by, or projected through, social media are alarming. Cases have erupted around the world that are leading to tremendous concern. Consider the case of the sinister Blue Whale social media ‘game’, which led to more than 130 teenage deaths in Russia.
An academic study on Chinese social media found that, in the broadcasting of a suicidal message, whereas a minority of observers did respond with positive messages, many chose to remain passive viewers. In Canada, the now infamous case of a repeatedly bullied girl, Amanda Todd, drew national attention after she broadcasted her suicide on Youtube.
Facebook’s new propositions include:
This builds on existing mechanisms that Facebook has put in place for suicide prevention, which were developed in collaboration with mental health organizations such as Save.org, National Suicide Prevention Lifeline, Forefront and Crisis Text Line.
The new tools depend on three important elements: partnerships with key mental health organizations, proactive user engagement, and Artificial Intelligence methods for filtering and reporting suicides.
While Facebook’s move should be seen in a positive light, given its power to save the lives of vulnerable people, it is also important to understand the balance that must be struck between safety and privacy.
This is by no means an easy tightrope to walk. On one hand, greater oversight of messages with potentially suicidal insinuations requires a greater intrusion into user outputs.
On the other hand, the early flagging of suicidal tendencies, which on social media have a statistically robust correspondence to actual self-harm, can protect people from serious damage.
So long as a balance is carefully set between safety and privacy, these new tools for suicide prevention may indeed prove life-saving for users who find themselves most alone and most vulnerable.