Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
Jump to content

2021 Facebook leak

From Wikipedia, the free encyclopedia
(Redirected from Facebook Papers)

In 2021, an internal document leak from the company then known as Facebook (now Meta Platforms, or Meta) showed it was aware of harmful societal effects from its platforms, yet persisted in prioritizing profit over addressing these harms. The leak, released by whistleblower Frances Haugen, resulted in reporting from The Wall Street Journal in September, as The Facebook Files series, as well as the Facebook Papers, by a consortium of news outlets the next month.

Primarily, the reports revealed that, based on internally-commissioned studies, the company was fully aware of negative impacts on teenage users of Instagram, and the contribution of Facebook activity to violence in developing countries. Other takeaways of the leak include the impact of the company's platforms on spreading false information, and Facebook's policy of promoting inflammatory posts. Furthermore, Facebook was fully aware that harmful content was being pushed through Facebook algorithms reaching young users. The types of content included posts promoting anorexia nervosa and self-harm photos.

In October 2021, Whistleblower Aid filed eight anonymous whistleblower complaints with the U.S. Securities and Exchange Commission (SEC) on behalf of Haugen alleging securities fraud by the company, after Haugen leaked the company documents the previous month.[1][2][3] After publicly revealing her identity on 60 Minutes,[4][5] Haugen testified before the U.S. Senate Commerce Subcommittee on Consumer Protection, Product Safety, and Data Security about the content of the leaked documents and the complaints.[6] After the company renamed itself as Meta Platforms,[7] Whistleblower Aid filed two additional securities fraud complaints with the SEC against the company on behalf of Haugen in February 2022.[8]

Background

[edit]

There were conflicts of interest between what was good for the public and what was good for Facebook. And Facebook, over and over again, chose to optimize for its own interests, like making more money.

Whistleblower Frances Haugen on 60 Minutes, October 3, 2021

In mid September 2021, The Wall Street Journal began publishing articles on Facebook based on internal documents from unknown provenance. Revelations included reporting of special allowances on posts from high-profile users ("XCheck"), subdued responses to flagged information on human traffickers and drug cartels, a shareholder lawsuit concerning the cost of Facebook (now Meta) CEO Mark Zuckerberg's personal liability protection in resolving the Cambridge Analytica data scandal, an initiative to increase pro-Facebook news within user news feeds, and internal knowledge of how Instagram exacerbated negative self-image in surveyed teenage girls.[9]

Siva Vaidhyanathan wrote for The Guardian that the documents were from a team at Facebook "devoted to social science and data analytics that is supposed to help the company's leaders understand the consequences of their policies and technological designs."[10] Casey Newton of The Verge wrote that it is the company's biggest challenge since its Cambridge Analytica data scandal.[11]

The leaked documents include internal research from Facebook that studied the impact of Instagram on teenage mental health.[12] Although Facebook claimed earlier that its rules applies equally to everyone on the platform, internal documents shared with The Wall Street Journal point to special policy exceptions reserved for VIP users, including celebrities and politicians.[13] After this reporting, Facebook's oversight board said it would review the system.[14][15]

On October 3, 2021, the former Facebook employee behind the leak, Frances Haugen, revealed her identity on 60 Minutes.[16]

The reports

[edit]

Beginning October 22, a group of news outlets began publishing articles based on documents provided by Haugen's lawyers, collectively referred to as The Facebook Papers.[17][18]

2020 U.S. elections and January 6 U.S. Capitol attack

[edit]

The New York Times pointed to internal discussions where employees raised concerns that Facebook was spreading content about the QAnon conspiracy theory more than a year before the 2020 United States elections. After the election, a data scientist mentioned in an internal note that 10 percent of all U.S. views of political content were of posts alleging that the election was fraudulent.[19] Among the ten anonymous whistleblower complaints Whistleblower Aid filed with the SEC on behalf of Haugen, one complaint alleged that Facebook misled the company's investors and the general public about its role in perpetuating misinformation related to the 2020 elections and political extremism that caused the January 6 United States Capitol attack.[1][4] Haugen was employed at Facebook from June 2019 until May 2021, starting within the company's Civic Integrity Team that was focused on investigating and addressing worldwide elections issues on the platform, as well as how the platform could be used to spread political disinformation and misinformation, to incite violence, and be abused by malicious governments until the company dissolved the team in December 2020.[20][21]

In the weeks after the 2020 U.S. presidential election, Facebook began rolling back many content policy enforcement measures it had in place during the election despite internal company tracking data showing a rise in policy-violating content on the platform, while Donald Trump's Facebook account had been whitelisted in the company's XCheck program.[21][22] Another of the whistleblower complaints Haugen filed with the SEC alleged that the company misled investors and the general public about enforcement of its terms of service due to such whitelisting under the XCheck program.[1][4] Haugen was interviewed by videoconference by the U.S. House Select Committee on the January 6 Attack in November 2021 about her tenure at Facebook, the company documents she provided to Congress, the company's corporate structure, and her testimony before Congress the previous month, but none of the information she provided to the Committee was included in its final report.[23][24]

Instagram's effects on teenagers

[edit]

The Files show that Facebook (now Meta) had been conducting internal research of how Instagram affects young users since 2018. While the findings point to Instagram being harmful to a large portion of young users, teenage girls were among the most harmed. Researchers within the company reported that "we make body issues worse for one in three teenage girls". Furthermore, internal research revealed that teen boys were also affected by negative social comparison, citing 14% of boys in the US in 2019.[25] Instagram was concluded to contribute to problems more specific to its app use, such as social comparison among teens.[26] Facebook published some of its internal research on September 29, 2021, saying these reports mischaracterized the purpose and results of its research.[27]

Studying of preteens

[edit]

The Files show that Facebook formed a team to study preteens, set a three year goal to create more products for this demographic, and commissioned strategy papers about the long-term business prospects of attracting the preteen demographic. A 2020 document from Facebook states: "Why do we care about tweens?" and answers that question by saying that "They are a valuable but untapped audience."[28]

Violence in developing countries

[edit]

An internal memo seen by the Washington Post revealed that Facebook has been aware of hate speech and calls for violence against groups like Muslims and Kashmiris, including posts of photos of piles of dead Kashmiri bodies with glorifying captions on its platform in India. Still, none of their publishers were blocked.[29] Documents reveal Facebook has responded to these incidents by removing posts which violate their policy, but has not made any substantial efforts to prevent repeat offenses.[29] As 90% of monthly Facebook users are now located outside of the US and Canada, Facebook claims language barriers are one obstacle that is preventing widespread reform.

Promoting anger-provoking posts

[edit]

In 2015, in addition to the Like button on posts, Facebook introduced a set of other emotional reaction options: love, haha, yay, wow, sad and angry.[30] The Washington Post reported that for three years, Facebook's algorithms promoted posts that received the 'angry' reaction from its users, based on internal analysis showing that such posts lead to five times more engagement than posts with regular likes. Years later, Facebook's researchers pointed out that posts with 'angry' reactions were much more likely to be toxic, polarizing, fake or low quality.[31]

In 2018, Facebook overhauled its News Feed algorithm, implementing a new algorithm which favored "Meaningful Social Interations" or "MSI". The new algorithm increased the weight of reshared material - a move which aimed to "reverse the decline in comments and encourage more original posting". While the algorithm was successful in its efforts, consequences such as user reports of feed quality decreasing along with increased anger on the site were observed. Leaked documents reveal that employees presented several potential changes to fix some of the highlighted issues with their algorithm. However, documents claim Mark Zuckerberg denied the proposed changes due to his worry that they might cause fewer users to engage with Facebook. Documents have also pointed to another 2019 study conducted by Facebook where a fake account based in India was created and studied to see what type of content it was presented and interacted with. Results of the study showed that within three weeks, the fake account's newsfeed was being presented pornography and "filled with polarizing and graphic content, hate speech and misinformation", according to an internal company report.[32]

Employee dissatisfaction

[edit]

Politico quotes several Facebook staff expressing concerns about the company's willingness and ability to respond to damage caused by the platform. A 2020 post reads: "It's not normal for a large number of people in the 'make the site safe' team to leave saying, 'hey, we're actively making the world worse FYI.' Every time this gets raised it gets shrugged off with 'hey people change jobs all the time' but this is NOT normal."[33]

Apple's threat to remove Facebook and Instagram

[edit]

In 2019, following concerns about Facebook and Instagram being used to trade maids in the Middle East, Apple threatened to remove their iOS apps from the App Store.[34]

XCheck

[edit]

The documents have shown a private program known as "XCheck" or "cross-check" that Facebook has employed in order to whitelist posts from users deemed as "high-profile". The system began as a quality control measure but has since grown to protect "millions of VIP users from the company's normal enforcement process". XCheck has led to celebrities and other public figures being exempt from punishment that the average Facebook user would receive from violating policies. In 2019, football player Neymar had posted nude photos of a woman who had accused him of rape which were left up for more than a day. According to The Wall Street Journal, "XCheck grew to include at least 5.8 million users in 2020" according to Facebook's internal documents.[35] The goal of XCheck was "to never publicly tangle with anyone who is influential enough to do you harm".[36]

Collaboration on censorship with the government of Vietnam

[edit]

In 2020, Vietnam's communist government threatened to shut down Facebook if the social media company did not cooperate on censoring political content in the country, Meta's (then known as Facebook) big market in Southeast Asia.[37] The decision to comply was personally approved by Mark Zuckerberg.[38][39]

Suppression of political movements on its platform

[edit]

In 2021, Facebook developed a new strategy for addressing harmful content on their site, implementing measures which were designed to reduce and suppress the spread of movements that were deemed hateful. According to a senior security official at Facebook, the company "would seek to disrupt on-platform movements only if there was compelling evidence that they were the product of tightly knit circles of users connected to real-world violence or other harm and committed to violating Facebook's rules". As part of their recently coordinated initiative, this included less promotion of the movement's posts within users' News Feed as well as not notifying users of new posts from these pages. Specific groups that have been highlighted as being affected by Facebook's social harm policy include the Patriot Party, previously linked to the Capitol attack, as well as a newer German conspiracy group known as Querdenken, who had been placed under surveillance by German intelligence after protests it organized repeatedly "resulted in violence and injuries to the police".[40]

Facebook's AI concern

[edit]

According to The Wall Street Journal, documents show that in 2019, Facebook reduced the time spent by human reviewers on hate-speech complaints, shifting towards a stronger dependence on their artificial intelligence systems to regulate the matter. However, internal documents from employees claim that their AI has been largely unsuccessful, seeing trouble detecting videos of cars crashing, cockfighting, as well as understanding hate speech in foreign languages.[41] Internal engineers and researchers within Facebook have estimated that their AI has only been able to detect and remove 0.6% of "all content that violated Facebook's policies against violence and incitement".[citation needed]

Inclusion of Breitbart News as trusted news source

[edit]

The Wall Street Journal reported that Facebook executives resisted removing the far-right website Breitbart News from Facebook's News Tab feature to avoid angering Donald Trump and Republican members of Congress, despite criticism from Facebook employees.[42][43] An August 2019 internal Facebook study had found that Breitbart News was the least trusted news source, and also ranked as low-quality, in the sources it looked at across the U.S. and Great Britain.[44]

The Wall Street Journal podcast

[edit]

For The Facebook Files series of reports, The Wall Street Journal produced a podcast on its The Journal channel, divided into eight episodes:

  • Part 1: The Whitelist[45]
  • Part 2: 'We Make Body Image Issues Worse'[46]
  • Part 3: 'This Shouldn't Happen on Facebook'[47]
  • Part 4: The Outrage Algorithm[48]
  • Part 5: The Push To Attract Younger Users[49]
  • Part 6: The Whistleblower[50]
  • Part 7: The AI Challenge[51]
  • Part 8: A New Enforcement Strategy[52]

Facebook's response

[edit]

In the Q3 2021 earnings call, Facebook CEO Mark Zuckerberg discussed the recent leaks, characterizing them as coordinated efforts to paint a false picture of his company by selectively leaking documents.[53]

According to a leaked internal email seen by The New York Times, Facebook asked its employees to "preserve internal documents and communications since 2016", a practice called a legal hold. The email continues: "As is often the case following this kind of reporting, a number of inquiries from governments and legislative bodies have been launched into the company's operations."[54]

Lobbying

[edit]

In December 2021, news broke on The Wall Street Journal pointing to Meta's lobbying efforts to divide US lawmakers and "muddy the waters" in Congress, to hinder regulation following the 2021 whistleblower leaks.[55] Facebook's lobbyist team in Washington suggested to Republican lawmakers that the whistleblower "was trying to help Democrats," while the narrative told to Democratic staffers was that Republicans "were focused on the company's decision to ban expressions of support for Kyle Rittenhouse," The Wall Street Journal reported. According to the article, the company's goal was to "muddy the waters, divide lawmakers along partisan lines and forestall a cross-party alliance" against Facebook (now Meta) in Congress.[56]

See also

[edit]

References

[edit]
  1. ^ a b c Zubrow, Keith; Gavrilovic, Maria; Ortiz, Alex (October 4, 2021). "Whistleblower's SEC complaint: Facebook knew platform was used to "promote human trafficking and domestic servitude"". 60 Minutes Overtime. CBS News. Archived from the original on May 12, 2022. Retrieved May 12, 2022.
  2. ^ Horwitz, Jeff (September 13, 2021). "Facebook Says Its Rules Apply to All. Company Documents Reveal a Secret Elite That's Exempt". The Wall Street Journal. News Corp. Archived from the original on September 5, 2023. Retrieved May 23, 2022.
  3. ^ Bursztynsky, Jessica; Feiner, Lauren (September 14, 2021). "Facebook documents show how toxic Instagram is for teens, Wall Street Journal reports". CNBC. Archived from the original on May 15, 2022. Retrieved May 12, 2022.
  4. ^ a b c Pelley, Scott (October 4, 2021). "Whistleblower: Facebook is misleading the public on progress against hate speech, violence, misinformation". 60 Minutes. CBS News. Archived from the original on October 4, 2021. Retrieved May 16, 2022.
  5. ^ Feiner, Lauren (October 3, 2021). "Facebook whistleblower reveals identity, accuses the platform of a 'betrayal of democracy'". CNBC. Archived from the original on May 17, 2022. Retrieved May 16, 2022.
  6. ^ Feiner, Lauren (October 5, 2021). "Facebook whistleblower: The company knows it's harming people and the buck stops with Zuckerberg". CNBC. Archived from the original on May 17, 2022. Retrieved May 16, 2022.
  7. ^ Bidar, Musadiq (October 28, 2021). "Facebook to change corporate name to Meta". CBS News. Archived from the original on May 17, 2022. Retrieved May 16, 2022.
  8. ^ Zakrzewski, Cat (February 18, 2022). "Facebook whistleblower alleges executives misled investors about climate, covid hoaxes in new SEC complaint". The Washington Post. Archived from the original on March 29, 2022. Retrieved May 12, 2022.
  9. ^ "Facebook Files: 5 things leaked documents reveal". BBC News. September 24, 2021. Archived from the original on October 4, 2021. Retrieved October 4, 2021.
  10. ^ Vaidhyanathan, Siva (October 8, 2021). "Facebook has just suffered its most devastating PR catastrophe yet". The Guardian. Archived from the original on October 9, 2021. Retrieved October 8, 2021.
  11. ^ Newton, Casey (September 28, 2021). "Why Facebook should release the Facebook Files". The Verge. Archived from the original on October 2, 2021. Retrieved October 4, 2021.
  12. ^ Gayle, Damien (September 14, 2021). "Facebook aware of Instagram's harmful effect on teenage girls, leak reveals". The Guardian. Archived from the original on October 10, 2021. Retrieved October 10, 2021.
  13. ^ Horwitz, Jeff (September 13, 2021). "Facebook Says Its Rules Apply to All. Company Documents Reveal a Secret Elite That's Exempt". Wall Street Journal. Archived from the original on September 5, 2023. Retrieved October 28, 2021.
  14. ^ "Facebook oversight board reviewing 'XCheck' system for VIPs". Associated Press. September 22, 2021. Archived from the original on October 28, 2021. Retrieved October 28, 2021.
  15. ^ "Facebook oversight board reviewing 'XCheck' system for VIPs". Associated Press. September 21, 2021. Archived from the original on October 28, 2021. Retrieved October 28, 2021.
  16. ^ Ghaffary, Shirin (October 3, 2021). "Why this Facebook scandal is different". Vox. Archived from the original on October 23, 2021. Retrieved October 24, 2021.
  17. ^ Danner, Chas (October 23, 2021). "What Was Leaked in the Facebook Papers?". Intelligencer. Archived from the original on October 28, 2021. Retrieved October 24, 2021.
  18. ^ Varnham O'Regan, Sylvia; Di Stefano, Mark (October 22, 2021). "New Facebook Storm Nears as CNN, Fox Business and Other Outlets Team Up on Whistleblower Docs". The Information. Archived from the original on October 25, 2021. Retrieved October 25, 2021.
  19. ^ Ryan Mac; Sheera Frenkel (October 22, 2021). "Internal Alarm, Public Shrugs: Facebook's Employees Dissect Its Election Role". The New York Times. Archived from the original on November 3, 2021. Retrieved October 29, 2021.
  20. ^ Horwitz, Jeff (October 3, 2021). "The Facebook Whistleblower, Frances Haugen, Says She Wants to Fix the Company, Not Harm It". The Wall Street Journal. News Corp. Archived from the original on November 3, 2021. Retrieved July 28, 2023.
  21. ^ a b Timberg, Craig; Dwoskin, Elizabeth; Albergotti, Reed (October 22, 2021). "Inside Facebook, Jan. 6 violence fueled anger, regret over missed warning signs". The Washington Post. Archived from the original on November 1, 2021. Retrieved July 28, 2023.
  22. ^ Horwitz, Jeff (September 13, 2021). "Facebook Says Its Rules Apply to All. Company Documents Reveal a Secret Elite That's Exempt". The Wall Street Journal. News Corp. Archived from the original on September 5, 2023. Retrieved July 28, 2023.
  23. ^ Zakrzewski, Cat; Lima, Cristiano; Harwell, Drew (January 17, 2023). "What the Jan. 6 probe found out about social media, but didn't report". The Washington Post. Archived from the original on August 3, 2023. Retrieved July 28, 2023.
  24. ^ Haugen, Frances (November 22, 2021). "Select Committee to Investigate the January 6th Attack on the U.S. Capitol, U.S. House of Representatives, Washington, D.C. – Interview of: Frances Haugen" (PDF) (Interview). Interviewed by U.S. House Select Committee on the January 6 Attack. Washington, D.C.: U.S. Government Publishing Office. Archived (PDF) from the original on July 28, 2023. Retrieved July 28, 2023.
  25. ^ "Facebook documents show how toxic Instagram is for teens, Wall Street Journal reports". CNBC. September 14, 2021. Archived from the original on November 11, 2021. Retrieved November 11, 2021.
  26. ^ "Facebook Knows Instagram is Toxic for Teen Girls, Company Documents Show". Wall Street Journal. September 14, 2021. Archived from the original on November 12, 2021. Retrieved November 11, 2021.
  27. ^ Musil, Steven (September 30, 2021). "Facebook releases internal research on Instagram's effects on teens, ahead of testimony". CNET. Archived from the original on July 22, 2022. Retrieved July 22, 2022.
  28. ^ Wells, Georgia; Horwitz, Jeff (September 28, 2021). "Facebook's Effort to Attract Preteens Goes Beyond Instagram Kids, Documents Show". The Wall Street Journal. ISSN 0099-9660. Archived from the original on April 12, 2022. Retrieved April 12, 2022.
  29. ^ a b Zakrzewski, Cat; De Vynck, Gerrit; Masih, Niha; Mahtani, Shibani (October 24, 2021). "How Facebook neglected the rest of the world, fueling hate speech and violence in India". The Washington Post. Archived from the original on October 30, 2021. Retrieved October 29, 2021.
  30. ^ "There is a specific sociological reason why Facebook introduced its new emoji 'reactions'". Business Insider. Archived from the original on October 30, 2021. Retrieved October 30, 2021.
  31. ^ "Facebook prioritized 'angry' emoji reaction posts in news feeds". The Washington Post. Archived from the original on October 30, 2021. Retrieved October 30, 2021.
  32. ^ Rai, Saritha (October 24, 2021). "In Just 21 Days, Facebook Led New India User to Porn, Fake News". Bloomberg News. Archived from the original on April 8, 2022. Retrieved April 10, 2022.
  33. ^ Hendel, John (October 25, 2021). "'This is NOT normal': Facebook employees vent their anguish". Politico. Archived from the original on October 27, 2021. Retrieved October 29, 2021.
  34. ^ Jon Gambrell; Jim Gomez (October 25, 2021). "Apple once threatened Facebook ban over Mideast maid abuse". AP. Archived from the original on October 27, 2021. Retrieved October 29, 2021.
  35. ^ Horwitz, Jeff (September 13, 2021). "Facebook Says Its Rules Apply to All. Company Documents Reveal a Secret Elite That's Exempt". Wall Street Journal. Archived from the original on November 3, 2021. Retrieved November 3, 2021.
  36. ^ Chappell, Bill (October 25, 2021). "The Facebook Papers: What you need to know about the trove of insider documents". NPR. Archived from the original on October 28, 2021. Retrieved October 29, 2021.
  37. ^ "Exclusive: Vietnam threatens to shut down Facebook over censorship requests - source". Reuters. November 20, 2020. Archived from the original on November 10, 2021. Retrieved November 3, 2021.
  38. ^ "Mark Zuckerberg was more involved in decision making at Facebook than he let on". The Washington Post. Archived from the original on October 30, 2021. Retrieved November 3, 2021.
  39. ^ "Opinion | Mark Zuckerberg is for free speech when it's convenient". MSNBC. October 25, 2021. Archived from the original on November 3, 2021. Retrieved November 3, 2021.
  40. ^ "Facebook Increasingly Suppresses Political Movements It Deems Dangerous". Wall Street Journal. October 22, 2021. Archived from the original on November 12, 2021. Retrieved November 11, 2021.
  41. ^ "Facebook Says AI Will Clean up the Platform. Its Own Engineers Have Doubts". Wall Street Journal. October 17, 2021. Archived from the original on November 12, 2021. Retrieved November 11, 2021.
  42. ^ Hagey, Keach; Horwitz, Jeff (October 24, 2021). "Facebook's Internal Chat Boards Show Politics Often at Center of Decision Making". The Wall Street Journal. ISSN 0099-9660. Archived from the original on December 3, 2021. Retrieved December 11, 2021.
  43. ^ Feinberg, Andrew (October 25, 2021). "Facebook protected Breitbart to avoid angering Trump, new leaks reveal". The Independent. Archived from the original on October 25, 2021. Retrieved December 11, 2021.
  44. ^ Hagey, Keach; Horwitz, Jeff (October 24, 2021). "Facebook's Internal Chat Boards Show Politics Often at Center of Decision Making". The Wall Street Journal. ISSN 0099-9660. Archived from the original on December 3, 2021. Retrieved December 11, 2021.
  45. ^ "The Facebook Files, Part 1: The Whitelist - The Journal. - WSJ Podcasts". WSJ. Archived from the original on December 21, 2021. Retrieved December 22, 2021.
  46. ^ "The Facebook Files, Part 2: 'We Make Body Image Issues Worse' - The Journal. - WSJ Podcasts". WSJ. Archived from the original on December 21, 2021. Retrieved December 22, 2021.
  47. ^ "The Facebook Files, Part 3: 'This Shouldn't Happen on Facebook' - The Journal. - WSJ Podcasts". WSJ. Archived from the original on December 21, 2021. Retrieved December 22, 2021.
  48. ^ "The Facebook Files, Part 4: The Outrage Algorithm - The Journal. - WSJ Podcasts". WSJ. Archived from the original on December 21, 2021. Retrieved December 22, 2021.
  49. ^ "The Facebook Files, Part 5: The Push To Attract Younger Users - The Journal. - WSJ Podcasts". WSJ. Archived from the original on December 21, 2021. Retrieved December 22, 2021.
  50. ^ "The Facebook Files, Part 6: The Whistleblower - The Journal. - WSJ Podcasts". WSJ. Archived from the original on December 21, 2021. Retrieved December 22, 2021.
  51. ^ "The Facebook Files, Part 7: The AI Challenge - The Journal. - WSJ Podcasts". WSJ. Archived from the original on December 21, 2021. Retrieved December 22, 2021.
  52. ^ "The Facebook Files, Part 8: A New Enforcement Strategy - The Journal. - WSJ Podcasts". WSJ. Archived from the original on December 21, 2021. Retrieved December 22, 2021.
  53. ^ "Third Quarter 2021 Results Conference Call" (PDF). Archived (PDF) from the original on October 26, 2021. Retrieved October 27, 2021.
  54. ^ Mac, Ryan; Isaac, Mike (October 27, 2021). "Facebook tells employees to preserve all communications for legal reasons". The New York Times. Archived from the original on October 28, 2021. Retrieved October 28, 2021.
  55. ^ "Facebook's Pushback: Stem the Leaks, Spin the Politics, Don't Say Sorry". Wall Street Journal. December 29, 2021. Archived from the original on August 14, 2022. Retrieved December 30, 2021.
  56. ^ "Facebook reportedly told Republicans whistleblower was 'trying to help Democrats'". December 29, 2021. Archived from the original on October 3, 2022. Retrieved December 30, 2021.

Further reading

[edit]
[edit]