New Suicide Prevention Tools

    Building a Safer Community With New Suicide Prevention Tools:

    ssi-live-broadcaster-support 
    Suicide Prevention By Vanessa Callison-Burch, Product Manager, Jennifer Guadagno, Researcher, and Antigone Davis, Head of Global Safety

    To begin with, there is one death by suicide in the world every 40 seconds. Suicide is the second leading cause of death for 15-29 year olds. Facebook is in a unique position to help connect a person in distress with people who can support them. It’s part of our ongoing effort to help build a safe community on and off Facebook.

    Today we’re updating the tools and resources we offer to people who may be thinking of suicide. Also, support is offered to their concerned friends and family members. Integrated suicide prevention tools to help people in real time on Facebook Live chat support from crisis support organizations through Messenger Streamlined reporting for suicide, assisted by artificial intelligence.

    Facebook:

    Already on Facebook if someone posts something that makes you concerned about their well-being, you can reach out to them directly or report the post to us. We have teams working around the world, 24/7. They review reports that come in and prioritize the most serious reports like suicide. In addition, we provide people who have expressed suicidal thoughts with a number of support options. For example, we prompt people to reach out to a friend and even offer pre-populated text to make it easier for people to start a conversation. We also suggest contacting a help line and offer other tips and resources for people to help themselves in that moment.

    Suicide prevention tools have been available on Facebook for more than 10 years.  They were developed in collaboration with mental health organizations. For instance, more popular resources include Save.org, National Suicide Prevention Lifeline, Forefront and Crisis Text Line. Furthermore, with input from people who have personal experience thinking about or attempting suicide. In 2016 we expanded the availability of the latest tools globally. With the help of over 70 partners around the world, we improved how they work based on new technology and feedback from the community.

    Supporting Someone on Facebook Live

    Our suicide prevention tools for Facebook posts will now be integrated into Facebook Live. People watching a live video have the option to reach out to the person directly and to report the video to us. We will also provide resources to the person reporting the live video. This assist them in helping their friend. The person sharing a live video will see a set of resources on their screen. They can choose to reach out to a friend, contact a help line or see tips. If you or someone you know is in crisis, it is important to call local emergency services right away. You can also visit our Help Center for information about how to support yourself or a friend.
    ssi-live-reporter-support

    Empowering Crisis Support Partners

    Partners are key to our work in suicide prevention and mental health support.

    In addition, we recently added the ability for people to connect with our crisis support partners over Messenger. Now people will see the option to message with someone in real time directly from the organization’s Page or through our suicide prevention tools. Participating organizations include Crisis Text Line, the National Eating Disorder Association and the National Suicide Prevention Lifeline. This test will expand over the next several months ensuring the organizations can support any new volume of communication. Zendesk donated some of the company’s backend tools to make this integration possible.

    Subsequently, we are also launching a video campaign with partner organizations across the globe. The goal is to raise awareness about ways to help a friend in need. Making Reporting Easier We work to address posts expressing thoughts of suicide as quickly and accurately as possible.

    Based on feedback from experts, we are testing a streamlined reporting process using pattern recognition in posts previously reported for suicide. This artificial intelligence approach will make the option to report a post about “suicide or self injury” more prominent for potentially concerning posts like these. We’re also testing pattern recognition to identify posts as very likely to include thoughts of suicide. Our Community Operations team will review these posts and, if appropriate, provide resources to the person who posted the content, even if someone on Facebook has not reported it yet.

    We are starting this limited test in the US and will continue working closely with suicide prevention experts to understand other ways we can use technology to help provide support. Suicide prevention is one way we’re working to build a safer community on Facebook. With the help of our partners and people’s friends and family members on Facebook, we’re hopeful we can support more people over time.