Analysis of Social Media Platform Mechanisms to Moderate Disinformation and Fake News; Facebook and Twitte

Document Type : Original Article

Author

Assistant Professor Department of Virtual Space Studies , Research Center For Culture, Art And Communications, Tehran, Iran.

Abstract

This article aims to analyze the policy and mechanisms of global social media platforms to deal with fake news and disinformation. This process is complex and challenging and is done with the help of human and machine agents. In this article, the steps and methods of content moderation on Facebook and Twitter platforms have been analyzed in a documentary way to identify their mechanisms to deal with fake news content. The findings show that these two platforms, although they operate in the same political and economic context and under the same media policy, have different mechanisms to deal with false information, which originates from their general approach. Facebook’s general approach is to defend transparency and give voice to people. Therefore, instead of deleting it, it has taken the reduction of the visibility of this type of content. Still, Twitter, to ensure the health of the conversation, removes fake content more easily. Finally, it must be said that the corporatization of dealing with false information and fake news makes it more difficult for citizens to receive real news, and the public sphere and democracy face increasingly complex risks. On the other hand, the dominance of companies over monitoring news content makes governments like the Iranian government face new challenges from the point of view of internet governance.

Keywords

About Twitter. (2021). “Our company”. Twitter website. https://about.twitter.com/en/who-we-are/our-company.
Bucher, Taina. (2021). Facebook. Cambridge: Polity.
Chan, K. and Casert, R. (2020). “EU law targets Big Tech over hate speech, disinformation”. ABC NEWS. https://abcnews.go.com/Business/wireStory/eu-law-targets-big-tech-hate-speech-disinformaion-84261130.
Conger, K. (2020). “Twitter has labeled 38% of Trump’s tweets since tuesday”. Nytimes.com.https://www.nytimes.com/2020/11/05/technology/donald-trump-twitter.html.
US Congress. (2022). Constitution of the United States. constitution.congress.gov.https://constitution.congress.gov/constitution/amendment -1/.
EFF. (2022). CDA 230: The Most Important Law Protecting Internet Speech. EFF.org.       https://www.eff.org/issues/cda230#:~:text=Section%20230%20says%20that%20%22No,%C2%A7%20230).
Facebook. (2021). “Facebook Community Standards: False News Policy Rationale”. Facebook Transparency Center. https://transparency.fb.com/en-gb/policies/community-  standards/false-news/.
Gillespie, T. (2018). Custodians of The Iinternet: platforms, content moderation, and the hidden decisions that shape social media. New Haven and London: Yale University Press.
Giusti, S. and Piras, E. (2021). Introduction: In search of paradigms: Disinformation, fake news, and post-truth politics; Democracy and Fake news: Information Manipulation and Post-Truth Politics. Edited by Serena Giusti and Elisa Piras; edited by Serena Giusti and Elisa        Piras, London and Paris: Routledge.
Grafanaki, S. (2018). “Platforms, the First Amendment and Online Speech: Regulating the Filters”. Pace Law Review. 39(1).
Mcintyre, L. (2018). Post-truth. Cambridge and London: MIT Press.
Mossery, A. (2017). “Working to Stop Misinformation and False News”, MetaWebsite.https://www.facebook.com/formedia/blog/working- to-stop-misinformation-and-false-news.
Murthy, D. (2018). Twitter: Social Communication in the Twitter Age. Cambridge and Medford: Polity.
Poell, T. et al. (2021). Platforms and Cultural Production. Cambridge and Medford: Polity.
Singh, S. (2019). “Everything in Moderation: An Analysis of How Internet Platforms Are Using       Artificial Intelligence to Moderate UserGenerated Content”. Open Technology Institute. https://www.newamerica.org/oti/reports/everything-moderation-analysis-how-internet-platforms-are-using-artificial-intelligence-moderate-user-generated-co ntent/the-limitations-of-automated-tools-in-content-moderation.
Twitter blog. (2020). “Protecting the public conversation”. https://blog.twitter.com/en_us/topics/company/2020/covid-19#protecting.
Twitter Help Center. (2022a). “Appeal an Account Suspension or Locked Account”. Twitter.com. https://help.twitter.com/forms/general?subtopic=suspended.
Twitter Help Center. (2022b). “How We Address Misinformation on Twitter”.Twitter.com. https://help.twitter.com/en/resources/addressing- misleading-info.
Twitter Help Center. (2022c). “Our Range of Enforcement Options”. Twitter.com. https://help.twitter.com/en/rules-and-policies/enforcement -options.
Wardle, C. and Derakhshan, H. (2017). Information Disorder: Toward an interdisciplinary  framework for research and policymaking. Council of Europe. file:///C:/Users/h.hasani/Downloads/162317GBR_Report%20desinformation%20(1).pdf.Roberts, S. T. (2017). Content Moderation; in Encyclopedia of Big Data. edited by Laurie A.   Schintler, Connie L. McNeely, Switzerland: Springer International Publishing.
Volume 4, Issue 7 - Serial Number 7
October 2022
Pages 155-184
  • Receive Date: 03 July 2022
  • Revise Date: 24 July 2022
  • Accept Date: 01 August 2022