Facebook is in the hot seat again.
The Wall Street Journal published a powerful multi-part series on the company this week, drawing from internal documents on everything from the company’s secretive practice of whitelisting celebrities to its knowledge that Instagram is taking a serious toll on the mental health of teen girls.
The flurry of investigative pieces makes it clear that what Facebook says in public doesn’t always reflect the company’s knowledge on known issues behind the scenes. The revelations still managed to shock even though Facebook has been playing dumb about the various social ills it has sown for years. (Remember when Mark Zuckerberg dismissed the notion that Facebook influenced the 2016 election as “crazy?”) Facebook’s longstanding PR playbook is to hide its dangers, denying knowledge of its darker impacts on society publicly, even as research spells them out internally.
That’s all well and good until someone gets ahold of the internal research.
One of the biggest revelations from the WSJ’s report: The company knows that Instagram poses serious dangers to mental health in teenage girls. An internal research slide from 2019 acknowledged that “We make body image issues worse for one in three teen girls” — a shocking admission for a company charging ahead with plans to expand to even younger and more vulnerable age groups.
As recently as May, Instagram’s Adam Mosseri dismissed concerns around the app’s negative impact on teens as “quite small.”
But internally, the picture told a different story. According to the WSJ, from 2019 to 2021, the company conducted a thorough deep dive into teen mental health, including online surveys, diary studies, focus groups and large-scale questionnaires.
According to one internal slide, the findings showed that 32% of teenage girls reported that Instagram made them have a worse body image. Of research participants who experienced suicidal thoughts, 13% of British teens and 6% of American teens directly linked their interest in killing themselves to Instagram.
“Teens blame Instagram for increases in the rate of anxiety and depression,” another internal slide stated. “This reaction was unprompted and consistent across all groups.”
Following the WSJ report, Senators Marsha Blackburn (R-TN) and Richard Blumenthal (D-CT) announced a probe into Facebook’s lack of transparency around internal research showing that Instagram poses serious and even lethal danger to teens. The Senate Subcommittee on Consumer Protection, Product Safety, and Data Security will launch the investigation.
“We are in touch with a Facebook whistleblower and will use every resource at our disposal to investigate what Facebook knew and when they knew it – including seeking further documents and pursuing witness testimony,” Senators Blackburn and Blumenthal wrote. “The Wall Street Journal’s blockbuster reporting may only be the tip of the iceberg.”
Blackburn and Blumenthal weren’t the only U.S. lawmakers alarmed by the new report. Sen. Ed Markey (D-MA), Rep. Kathy Castor (D-FL), and Lori Trahan (D-MA) sent Facebook their own letter demanding that the company walk away from its plan to launch Instagram for kids. “Children and teens are uniquely vulnerable populations online, and these findings paint a clear and devastating picture of Instagram as an app that poses significant threats to young people’s wellbeing,” the lawmakers wrote.
In May, a group of 44 state attorneys general wrote to Instagram to encourage the company to abandon its plans to bring Instagram to kids under the age of 13. “It appears that Facebook is not responding to a need, but instead creating one, as this platform appeals primarily to children who otherwise do not or would not have an Instagram account,” the group of attorneys general wrote. They warned that an Instagram for kids would be “harmful for myriad reasons.”
In April, a collection of the same Democratic lawmakers expressed “serious concerns” about Instagram’s potential impact on the well-being of young users. That same month, a coalition of consumer advocacy organizations also demanded that the company reconsider launching a version of Instagram for kids.
According to the documents obtained by the WSJ, all of those concerns look extremely valid. In spite of extensive internal research and their deeply troubling findings, Facebook has downplayed its knowledge publicly, even as regulators regularly pressed the company for what it really knows.
Instagram’s Mosseri may have made matters worse Thursday when he made a less than flattering analogy between social media platforms and vehicles. “We know that more people die than would otherwise because of car accidents, but by and large, cars create way more value in the world than they destroy,” Mosseri told Peter Kafka on Recode’s media podcast. “And I think social media is similar.”
Mosseri dismissed any comparison between social media and drugs or cigarettes in spite of social media’s well-researched addictive effects, likening social platforms to the auto industry instead. Naturally, the company’s many critics jumped on the car comparison, pointing to their widespread lethality and the fact that the auto industry is heavily regulated — unlike social media.
Kenya labor court rules that Facebook can be sued
NAIROBI, Kenya (AP) — A judge in Kenya has ruled that Facebook’s parent company, Meta, can be sued in the East African country.
Meta tried to have the case dropped, arguing that Kenyan courts do not have jurisdiction over their operations, but the labor court judge dismissed that in a ruling on Monday.
A former Facebook moderator in Kenya, Daniel Motaung, is suing the company claiming poor working conditions.
Motaung said that while working as a moderator he was exposed to gruesome content such as rape, torture and beheadings that risked his and colleagues’ mental health.
He said Meta did not offer mental health support to employees, required unreasonably long working hours, and offered minimal pay. Motaung worked in Facebook’s African hub in Kenya’s capital, Nairobi, which is operated by Samasource Ltd.
Following the judge’s decision that Meta can be sued in Kenya, the next step in case will be considered by the court on Mar. 8.
Meta is facing a separate court case in which two Ethiopians say hate speech was allowed and even promoted on Facebook amid heated rhetoric over their country’s deadly Tigray conflict.
That lawsuit alleges that Meta hasn’t hired enough content moderators to adequately monitor posts, that it uses an algorithm that prioritizes hateful content, and that it responds more slowly to crises in Africa than elsewhere in the world.
The Associated Press and more than a dozen other media outlets last year reported that Facebook had failed to quickly and effectively moderate hate speech in several places around the world, including in Ethiopia. The reports were based on internal Facebook documents leaked by former employee and whistleblower Frances Haugen.