Facebook has announced a set of new tools to help detect suicidal intent messages, live videos, and posts. It has released these tools as a response to a growing international public health problem: one person commits suicide every 40 seconds, according the World Health Organization (WHO) Report Preventing Suicide: A Global Imperative.
There are three reasons why social media is in a unique position to detect, as well as to help prevent, suicide and other grave self-harm.
- First, there is an age factor: among young people 15-29 years of age, suicide is the second leading cause of death globally, according to the WHO. This cohort is a core user base of social media.
- Second, the WHO notes that suicides occur in all regions of the world. This makes the networked systems of social media more efficient trackers of suicidal messages at a global level.
- Third, the WHO points out that most suicides occur in low- and middle-income countries where resources are limited for early identification of people in need. This is where Facebook's early identification can be highly suitable.
In addition, there exists a phenomenon of suicides on social media itself, whereby people have actually broadcast their own suicides to online audiences.
Incidences of mortal self-harm that are influenced by, or projected through, social media are alarming. Cases have erupted around the world that are leading to tremendous concern. Consider the case of the sinister Blue Whale social media 'game', which led to more than 130 teenage deaths in Russia.
An academic study on Chinese social media found that, in the broadcasting of a suicidal message, whereas a minority of observers did respond with positive messages, many chose to remain passive viewers. In Canada, the now infamous case of a repeatedly bullied girl, Amanda Todd, drew national attention after she broadcasted her suicide on Youtube.
Facebook's new propositions include:
- Integrated suicide prevention tools to help people in real time on Facebook Liv
- Live chat support from crisis support organizations through Messenger
- Streamlined reporting for suicide, assisted by artificial intelligence
This builds on existing mechanisms that Facebook has put in place for suicide prevention, which were developed in collaboration with mental health organizations such as Save.org, National Suicide Prevention Lifeline, Forefront and Crisis Text Line.
The new tools depend on three important elements: partnerships with key mental health organizations, proactive user engagement, and Artificial Intelligence methods for filtering and reporting suicides.
While Facebook's move should be seen in a positive light, given its power to save the lives of vulnerable people, it is also important to understand the balance that must be struck between safety and privacy.
This is by no means an easy tightrope to walk. On one hand, greater oversight of messages with potentially suicidal insinuations requires a greater intrusion into user outputs.
On the other hand, the early flagging of suicidal tendencies, which on social media have a statistically robust correspondence to actual self-harm, can protect people from serious damage.
So long as a balance is carefully set between safety and privacy, these new tools for suicide prevention may indeed prove life-saving for users who find themselves most alone and most vulnerable.