Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Automated Transparency: A Legal and Empirical Analysis of the Digital Services Act Transparency Database (2404.02894v2)

Published 3 Apr 2024 in cs.CY and cs.SI

Abstract: The Digital Services Act (DSA) is a much awaited platforms liability reform in the European Union that was adopted on 1 November 2022 with the ambition to set a global example in terms of accountability and transparency. Among other obligations, the DSA emphasizes the need for online platforms to report on their content moderation decisions (`statements of reasons' - SoRs), which is a novel transparency mechanism we refer to as automated transparency in this study. SoRs are currently made available in the DSA Transparency Database, launched by the European Commission in September 2023. The DSA Transparency Database marks a historical achievement in platform governance, and allows investigations about the actual transparency gains, both at structure level as well as at the level of platform compliance. This study aims to understand whether the Transparency Database helps the DSA to live up to its transparency promises. We use legal and empirical arguments to show that while there are some transparency gains, compliance remains problematic, as the current database structure allows for a lot of discretion from platforms in terms of transparency practices. In our empirical study, we analyze a representative sample of the Transparency Database (131m SoRs) submitted in November 2023, to characterise and evaluate platform content moderation practices.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (27)
  1. Robyn Caplan and Tarleton Gillespie. 2020. Tiered Governance and Demonetization: The Shifting Terms of Labor and Compensation in the Platform Economy. Social Media + Society 6, 2 (April 2020), 205630512093663. https://doi.org/10.1177/2056305120936636
  2. One Day in Content Moderation: Analyzing 24 h of Social Media Platforms’ Content Decisions through the DSA Transparency Database. Technical Report. Center for Media, Communication and Information Research, Bremen. https://platform-governance.org/2023/one-day-in-content-moderation-by-social-media-platforms-in-the-eu/
  3. Chiara Drolsbach and Nicolas Pröllochs. 2023. Content Moderation on Social Media in the EU: Insights From the DSA Transparency Database. http://arxiv.org/abs/2312.04431 arXiv:2312.04431 [cs].
  4. Access to data and algorithms. https://pure.uvt.nl/ws/portalfiles/portal/68417348/CERRE_DMA_DSA_Report_FINAL.pdf.
  5. European Commission. 2019. The EU Code of conduct on countering illegal hate speech online. https://commission.europa.eu/strategy-and-policy/policies/justice-and-fundamental-rights/combatting-discrimination/racism-and-xenophobia/eu-code-conduct-countering-illegal-hate-speech-online_en.
  6. European Commission. 2022. The 2022 Code of Practice on Disinformation. https://digital-strategy.ec.europa.eu/en/policies/code-practice-disinformation.
  7. Mikkel Flyverbom. 2016. Transparency: Mediation and the Management of Visibilities. International Journal of Communication 10 (2016), 110–122.
  8. Full Disclosure: The Perils and Promise of Transparency. Cambridge University Press. Google-Books-ID: vSJcyLQgrKcC.
  9. Tarleton Gillespie. 2018. Custodians of the Internet (1st ed.). Yale University Press, New Haven.
  10. The Case for a Legal Compliance API for the Enforcement of the EU’s Digital Services Act on Social Media Platforms. In Proceedings of the 2022 ACM Conference on Fairness, Accountability, and Transparency (Seoul, Republic of Korea) (FAccT ’22). Association for Computing Machinery, New York, NY, USA, 1341–1349. https://doi.org/10.1145/3531146.3533190
  11. Algorithmic content moderation: Technical and political challenges in the automation of platform governance. Big Data and Society 7, 1 (Jan. 2020), 205395171989794. https://doi.org/10.1177/2053951719897945
  12. Amélie Heldt. 2019. Reading between the lines and the numbers: an analysis of the first NetzDG reports. Internet Policy Review 8, 2 (June 2019). https://doi.org/10.14763/2019.2.1398
  13. Does Transparency in Moderation Really Matter? User Behavior After Content Removal Explanations on Reddit. Proc. ACM Hum.-Comput. Interact. 3, CSCW, Article 150 (nov 2019), 27 pages. https://doi.org/10.1145/3359252
  14. Through the Looking Glass: Study of Transparency in Reddit’s Moderation Practices. Proc. ACM Hum.-Comput. Interact. 4, GROUP, Article 17 (jan 2020), 35 pages. https://doi.org/10.1145/3375197
  15. Lisa Kosters and Oskar Josef Gstrein. 2024. TikTok and transparency obligations in the forthcoming EU Digital Services Act (DSA)–A scoping review. Forthcoming in Zeitschrift für Europarechtliche Studien (ZEuS) 1 (2024).
  16. Paddy Leerssen. 2023. An end to shadow banning? Transparency rights in the Digital Services Act between content moderation and curation. Computer Law and Security Review 48 (April 2023), 105790. https://doi.org/10.1016/j.clsr.2023.105790
  17. Kim Lyons. 2021. Facebook reportedly provided inaccurate data to misinformation researchers — theverge.com. https://www.theverge.com/2021/9/11/22668396/facebook-provided-inaccurate-data-misinformation-researchers.
  18. Sarah Myers West. 2018. Censored, suspended, shadowbanned: User interpretations of content moderation on social media platforms. New Media and Society 20, 11 (May 2018), 4366–4383. https://doi.org/10.1177/1461444818773059
  19. Brandie Nonnecke and Camille Carlton. 2022. EU and US legislation seek to open up digital platform data. Science 375, 6581 (Feb. 2022), 610–612. https://doi.org/10.1126/science.abl8537
  20. European Parliament. 2019. Directive 2019/790 on Copyright in the Digital Single Market. https://eur-lex.europa.eu/eli/dir/2019/790/oj.
  21. European Parliament. 2022. Regulation (EU) 2022/2065 of the European Parliament and of the Council of 19 October 2022 on a Single Market For Digital Services and amending Directive 2000/31/EC (Digital Services Act). https://eur-lex.europa.eu/legal-content/EN/TXT/?uri=celex%3A32022R2065.
  22. Frank Pasquale. 2016. The Black Box Society: the Secret Algorithms that Control Money and Information (first harvard university press paperback edition ed.). Harvard University Press, Cambridge, Massachusetts London, England.
  23. Ben Popper. 2015. Airbnb’s worst problems are confirmed by its own data — theverge.com. https://www.theverge.com/2015/12/4/9849242/airbnb-data-new-york-affordable-housing-illegal-hotels.
  24. Sarah T. Roberts. 2018. Digital detritus: “Error” and the logic of opacity in social media content moderation. First Monday (March 2018). https://doi.org/10.5210/fm.v23i3.8283
  25. Jan M. Smits. 2017. What Is Legal Doctrine?: On The Aims and Methods of Legal-Dogmatic Research. Cambridge University Press, 207–228. https://doi.org/10.1017/9781316442906.006
  26. What Do We Mean When We Talk About Transparency? Toward Meaningful Transparency in Commercial Content Moderation. International Journal of Communication 13 (2019), 1526–1543.
  27. The DSA Transparency Database: Auditing Self-reported Moderation Actions by Social Media. http://arxiv.org/abs/2312.10269 arXiv:2312.10269 [cs].
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Rishabh Kaushal (7 papers)
  2. Jacob van de Kerkhof (3 papers)
  3. Catalina Goanta (11 papers)
  4. Gerasimos Spanakis (51 papers)
  5. Adriana Iamnitchi (34 papers)
Citations (2)

Summary

Automated Transparency through the DSA Transparency Database: Legal and Empirical Insights

Introduction

The Digital Services Act (DSA) represents a significant regulatory milestone within the European Union, aiming to enhance accountability and transparency in digital platform governance. A critical component of the DSA is the obligation for online platforms to report content moderation decisions via Statements of Reasons (SoRs), which are compiled in the DSA Transparency Database. This novel mechanism of automated transparency allows for a detailed examination of platform compliance and content moderation practices, setting a new benchmark in platform governance. This blog post explores the effectiveness of the DSA Transparency Database in fulfilling the DSA's transparency objectives, addressing the compliance issues observed, and examining the potential implications for future developments in AI and regulatory practices.

The DSA Transparency Database

The DSA Transparency Database, initiated in September 2023, marks a pivotal step in the quest for greater transparency in online content moderation. It requires Very Large Online Platforms (VLOPs)—defined as platforms with at least 45 million monthly active users in the EU—to report their content moderation decisions. This database is not only aimed at fostering transparency but also serves as a tool for investigating the adherence of platforms to DSA mandates on a scale previously unseen. As of February 2024, all online platforms engaged in content moderation within the EU are obliged to contribute SoRs to the database, underpinning the extensive scope of this transparency initiative.

Compliance and Transparency Gains

A thorough examination of the database's schema against the prerequisites of Articles 17 and 24 DSA indicates a designed framework conducive to transparency. However, it also reveals the discretionary power left to platforms regarding the extent of information disclosed in their transparency practices. The empirical paper of submitted SoRs unveils varying degrees of compliance across platforms, reflecting a rich tapestry of content moderation landscapes. Despite some platforms embracing their reporting obligations, the paper identifies gaps pointing to selective transparency and potential non-compliance areas, underscoring the complexity of operationalizing DSA mandates.

Platform Practices and DSA Compliance

The analysis categorizes platform compliance and practices into who is reporting, how moderation decisions and detection are executed, why content is moderated, what moderation actions are reported, where content is moderated in terms of language and territorial scope, and when SoRs are submitted. Notably, the database's empirical data reveal a predominant reliance on Terms of Service (ToS) violations over illegal content as grounds for moderation. This preference underscores the nuanced balance platforms navigate between regulatory compliance and the autonomy of their governance policies.

Implications and Future Directions

The DSA Transparency Database initiates a critical dialogue on the efficacy of automated transparency in regulating platform governance. While the database showcases a stride toward transparency, it concurrently highlights the challenges in ensuring full compliance and the efficacy of content moderation disclosures. The findings suggest a need for more robust mechanisms to enhance the clarity, specificity, and meaningfulness of SoRs, ensuring that they serve their intended purpose of accountability.

Moreover, the dataset's insights call for ongoing research to understand better the compliance nuances and the broader implications of such transparency initiatives on content moderation practices. As the regulatory landscape continues to evolve, the learnings from the DSA Transparency Database offer valuable lessons for future AI and regulatory policy developments, emphasizing the need for adaptable, nuanced, and collaborative approaches to governing the digital ecosystem.

Conclusion

The DSA Transparency Database serves as a foundational step towards realizing the ambitious transparency goals of the DSA. While it unveils the diversity and complexity of platform content moderation practices, it also brings to light the challenges inherent in standardizing transparency across varied platforms. As the database continues to evolve, it will be imperative to refine its mechanisms, ensuring it effectively contributes to the broader objectives of fairness, accountability, and transparency in the digital space.