Developing a Content Moderation Framework for Fake News on Social Media (A Case Study of X and Instagram)

Document Type : Research Paper

Authors

1 Associate Prof., Department of Media Management, Iranian Research Institute for Information Science and Technology (IranDoc), Tehran, Iran.

2 MSc., Department of Media Management, School of Culture & Communication, Sooreh University, Tehran, Iran.

10.22059/jibm.2024.378826.4812

Abstract

Objective
In today’s digital landscape, the rapid spread of unverified and misleading information, particularly on social media, poses significant challenges to modern societies. Fake news has emerged as a critical issue, threatening public trust, fueling social polarization, and damaging the legitimacy of institutions. However, beyond the issue of fake news itself lies a more systemic and strategic problem: the necessity of effective content moderation. As social media platforms play an increasingly central role in shaping public opinion, distributing information, and facilitating discourse, the ability to moderate fake news within these platforms has emerged as a critical dimension of media governance. Content moderation in this context is not limited to deleting or labeling posts; rather, it encompasses a multi-layered process involving policymaking, technological tools, user engagement, and cultural adaptation. This study aims to develop a comprehensive, context-sensitive framework for moderating fake news on domestic platforms by examining the strategies employed by two global leaders in the field—X (formerly Twitter) and Instagram. Through comparative analysis and in-depth engagement with domain experts, the research attempts to localize and adapt these strategies to the specific socio-cultural and regulatory environment of Iran.
 
Methodology
This research adopts a qualitative, comparative approach. Initially, the study reviewed policy documents, moderation guidelines, and institutional reports from both platforms to identify their respective content moderation mechanisms. In the second phase, data were gathered through semi-structured interviews with experts in digital policy, media regulation, and information governance. Thematic analysis was used to code and synthesize findings in a structured and iterative manner, facilitating the development of a localized and actionable moderation framework.
 
Findings
The results indicate that effective fake news moderation requires a dual-layered approach involving both off-platform and on-platform interventions. Off-platform strategies include legislative development, promotion of reliable news sources, enhancement of media literacy, institutional empowerment, public education campaigns, fostering critical discourse, development of detection algorithms, and encouraging civic engagement. On-platform strategies encompass internal content moderation guidelines, user verification systems, crowdsourcing fact-checking tasks, collaborations with official news organizations, algorithmic filtering, labeling and ranking content, and enabling transparent user feedback mechanisms. Notably, Iranian experts emphasized the importance of aligning these measures with local legal and cultural norms, ensuring algorithmic transparency, and offering users clear paths for appeal and redress.
 
Conclusion
Fake news, amplified by viral digital dynamics, can erode social cohesion and destabilize public discourse. However, the deployment of structured and context-aware content moderation strategies can both limit the spread of false information and enhance user trust in digital environments. The framework proposed in this study synthesizes international practices with localized insights to deliver a hybrid model suitable for domestic social media platforms. This model not only serves as a policy guideline for platform developers and regulators but also contributes to the broader agenda of rebuilding public trust and ensuring the credibility of online information ecosystems in Iran and similar contexts.

Keywords

Main Subjects


 
Abreu, A. I. A., de Zúñiga Navajas, H. G., & McCombs, M. E. (2020). Orígenes y desarrollo de la teoría de la agenda setting en Comunicación. Tendencias en España (2014-2019). El profesional de la información29(4), 18.
Barberá, P., Jost, J. T., Nagler, J., Tucker, J. A., & Bonneau, R. (2015). Tweeting from left to right: Is online political communication more than an echo chamber?. Psychological science26(10), 1531-1542.
Barclay, D.A. (2018). Fake News, Propaganda, and Plain Old Lies. Rowman & Littlefield Publishers.
Brashier, N. M., Pennycook, G., Berinsky, A. J., & Rand, D. G. (2021). Timing matters when correcting fake news. Proceedings of the National Academy of Sciences118(5), e2020043118.
Calzada, J. & Gil, R. (2020). What do news aggregators do? Evidence from Google News in Spain and Germany. Marketing Science, 39(1), 134-167.
Carlson, M. (2018). Facebook in the News: social media, Journalism, and Public Responsibility following the 2016 Trending Topics controversy. Digital Journalism 6(1), 4-20.
Common, M. F. (2020). Fear the reaper: How content moderation rules are enforced on social media. International Review of Law, Computers & Technology34(2), 126-152.
Cueva, E., Ee, G., Iyer, A., Pereira, A., Roseman, A., & Martinez, D. (2020, October). Detecting fake news on twitter using machine learning models. In 2020 IEEE MIT undergraduate research technology conference (URTC) (pp. 1-5). IEEE.
Evans, R., Jackson, D. & Murphy, J. (2023). Google News and machine gatekeepers: Algorithmic personalisation and news diversity in online news search. Digital Journalism, 11(9), 1682-1700.
Gallo, J. A., & Cho, C. Y. (2021). Social media: Misinformation and content moderation issues for congress. Congressional Research Service Report46662.
Gaozhao, D. (2021). Flagging fake news on social media: An experimental study of media consumers' identification of fake news. Government Information Quarterly, 38(3), 101591. https://doi.org/10.1016/j.giq.2021.101591
Gillespie, T. (2018). Custodians of the Internet: Platforms, content moderation, and the hidden decisions that shape social media. Yale University Press.
Grygiel, & Brown, N. (2019). Are social media companies motivated to be good corporate citizens? Examination of the connection between corporate social responsibility and social media safety. Telecommunications policy, 43(5), 445-460.
Harambam, J., Hoboken, J. & Helberger, N. (2018). Democratizing algorithmic news recommenders: How to materialize voice in a technologically saturated media ecosystem. Philosophical Transactions of the Royal Society, 376(2133), 1–21. https://doi.org/10.1098/rsta.2018.0088
Hassani, H. (2022). Analysis of Social Media Platform Mechanism to Moderate Disinformation and Fake News; Facebook and Twitte. Art & Media Studies, 4(7), 155- 184. https://civilica.com/doc/1534985 (in Persian)
Heitz, L., Lischka, J. A., Birrer, A., Paudel, B., Tolmeijer, S., Laugwitz, L., & Bernstein, A. (2022). Benefits of diverse news recommendations for democracy: A user study. Digital Journalism, 10(10), 1710-1730. doi:10.1080/21670811.2021.2021804
Kwak, k, T., Lee, S. & Lee, S. W. (2021). News and user characteristics used by personalized algorithms: The case of Korea's news aggregators, Naver News and Kakao News. Technological Forecasting and Social Change, 171(2), 1-12.
Lewandowsky, S. & Van Der Linden, S. (2021). Countering misinformation and fake news through inoculation and prebunking. European review of social psychology, 32(2), 348-384.
Majdizade, Z. and Mousavi Haghshenas, M. (2020). The “Why & how” of Fake News in Digital Age: A Meta Analysis. Journal of Interdisciplinary Studies in Communication and Media, 3(7), 93-124. doi: 10.22034/jiscm.2020.222901.1099 (in Persian)
Møller, L. A. (2022). Between personal and public interest: How algorithmic news recommendation reconciles with journalism as an ideology. Digital journalism, 10(10), 1794-1812.
Nechushtai, E. & Lewis, S. C. (2019). What kind of news gatekeepers do we want machines to be? Filter bubbles, fragmentation, and the normative dimensions of algorithmic recommendations. Computers in human behavior, 90, 298-307
Rahmanian, E. (2021). Typology of Fake Social Media News in the Context of Iran: An Ethnographic Approach. Journal of Business Management, 13(3), 814-844. doi: 10.22059/jibm.2021.319197.4066 (in Persian)
Riedl, M. J., Whipple, K. N., & Wallace, R. (2022). Antecedents of support for social media content moderation and platform regulation: the role of presumed effects on self and others. Information, Communication & Society25(11), 1632-1649.
Saei, M. H. & Azadi, M. H. (2021). The solutions of combating with fake news in Iranian media and news organizations. Rasaneh, 32(2), 35-78. doi: 10.22034/bmsp.2021.136841
(in Persian)
Van der Linden, S., Roozenbeek, J., Maertens, R., Basol, M., Kácha, O., Rathje, S., & Traberg, C. S. (2021). How can psychological science help counter the spread of fake news?. The Spanish journal of psychology24, e25.
Vázquez-Herrero, J., Vizoso, Á., & López-García, X. (2019). Innovación tecnológica y comunicativa para combatir la desinformación: 135 experiencias para un cambio de rumbo. Profesional de la información28(3).
Wu, H. D. & Liu, B. F. (2018). The effects of content moderation on user engagement and community dynamics: Evidence from Reddit. Information Systems Research, 29(3), 631-646.