Molly Rose Russell was by all accounts a normal
14-year-old schoolgirl.
She had the interests and hobbies of a typical teenager: music from “Hamilton,” the band 5 Seconds of Summer, a lead role in her school play. A
“positive, happy, bright young lady who was indeed destined to do good,” explained her father, as reported in The Guardian. Molly’s social media activity – music,
fashion, jewelry, Harry Potter – was what you might expect of a teenage girl.
But slowly, something changed. Her father told her the family was concerned about her, but she
said simply it’s “just a phase I’m going through.” In weeks after that conversation, Molly and her family had dinner together and then sat down to watch TV in the living
room.
At 7 a.m. the next day, Molly’s mother opened the door to her bedroom and found her daughter’s body. Molly had taken her own life.
Teen suicide
isn’t something we usually talk about. The media resists reporting suicide, treating it often as a private matter. And for parents of teens -- I’ve had two of them -- it’s almost
impossible to determine the line between normal teen troubles and the downward spiral that results in heartbreaking despondency and death.
Stop here for a moment. Molly’s story is
different. And it shines a powerful light on what more parents and teens are dealing with than you probably know.
Molly died in November 2017. At a coroner’s inquest held last
week in London, the dark world that Molly had found herself in was laid out in stark relief, concluding “a legal battle that pitted the Russell family against some of Silicon Valley’s
largest companies,” according to The New York Times.
advertisement
advertisement
On Instagram, Molly viewed a stream of dark content, including videos related to suicide, depression and self-harm. In
total, Molly binged-watched 138 videos that contained suicide and self-harm content. Of 16,300 pieces of content Molly saved on Instagram in the six months before she died, 2,100 were related to
depression and suicide. She last used her iPhone to access Instagram on the day of her death, at 12:45 a.m.
Coroner Andrew Walker told the court, after conducting a thorough
investigation that social networks like Instagram and Pinterest were “not safe" and "shouldn't have been available for a child to see," as reported by BBC News. Concluding it would not be "safe"
to rule Molly's cause of death was suicide, Mr. Walker said the teenager "died from an act of self-harm while suffering depression and the negative effects of online content,"
Walker was clear
on who he held responsible for Molly’s death. "The platform operated in such a way using algorithms as to result, in some circumstances, of binge periods of images, video clips, and text - some
of which were selected and provided without Molly requesting them.”
These images weren't sought out. They were algorithmically calculated and served to the 14-year-old.
Princeton Professor of Psychology and Neuroscience Uri Hasson calls this behavior no different than a drug dealer. He says without ambiguity that social networks are addictive.
This
is a terrible story. And the coroner’s ruling comes after Francis Haugen’s important efforts to expose internal research on the harm Instagram could cause teen girls, and Tristan
Harris’s efforts to expose the internal efforts to amplify engagement at Facebook at any cost.
But Walker’s ruling connects two disturbing dots, showing that Instagram
is doing what Molly’s father Ian Russell called “monetizing misery.”
So let’s look at the data.
The Journal of Psychiatry reports in the
article “Social media, internet use and suicide attempts in adolescents” that “there is an independent association between problematic use of social media/internet and suicide
attempts in young people.”
The article goes on, "Themes such as self-loathing, loneliness, and feeling unloved were found in content analysis of 3360 randomly selected posts from 17
depression-related accounts on Tumblr. There is an association between comments on Instagram with increasing severity of self-injury, suggesting social media may act to reinforce harmful
behaviors.”
Meta, the corporate owner of Instagram, disputes these findings, saying “It is simply not accurate that this research demonstrates Instagram is
‘toxic’ for teen girls.”
But testifying before the UK coroner’s inquest, Meta’s Elizabeth Lagone was less clear.
“Do you think this
type of material is safe for children?” Oliver Sanders, the Russell family attorney, asked on cross-examination. “I think it is safe for people to be able to express themselves,”
replied Lagone. “So you are saying yes, it is safe or no, it isn’t safe?” asked Sanders. “Yes, it is safe,” Lagone replied.
But father Ian Russell doesn't
buy it. He says Instagram “helped kill my daughter.”
After Molly took her own life, the Russells began to look at what the Instagram algorithm had been sending
her.
One account Molly followed featured an image of a blindfolded girl hugging a teddy bear with the text: “This world is so cruel, and I don’t wanna see it any
more.”
The Russells say that Molly was sent self-harm images even after her death.