Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 170 tok/s
Gemini 2.5 Pro 50 tok/s Pro
GPT-5 Medium 30 tok/s Pro
GPT-5 High 41 tok/s Pro
GPT-4o 60 tok/s Pro
Kimi K2 208 tok/s Pro
GPT OSS 120B 440 tok/s Pro
Claude Sonnet 4.5 35 tok/s Pro
2000 character limit reached

Longitudinal Monitoring of LLM Content Moderation of Social Issues (2510.01255v1)

Published 24 Sep 2025 in cs.CL, cs.CY, and cs.HC

Abstract: LLMs' (LLMs') outputs are shaped by opaque and frequently-changing company content moderation policies and practices. LLM moderation often takes the form of refusal; models' refusal to produce text about certain topics both reflects company policy and subtly shapes public discourse. We introduce AI Watchman, a longitudinal auditing system to publicly measure and track LLM refusals over time, to provide transparency into an important and black-box aspect of LLMs. Using a dataset of over 400 social issues, we audit Open AI's moderation endpoint, GPT-4.1, and GPT-5, and DeepSeek (both in English and Chinese). We find evidence that changes in company policies, even those not publicly announced, can be detected by AI Watchman, and identify company- and model-specific differences in content moderation. We also qualitatively analyze and categorize different forms of refusal. This work contributes evidence for the value of longitudinal auditing of LLMs, and AI Watchman, one system for doing so.

Summary

  • The paper introduces AI Watchman, a system for longitudinally monitoring LLM refusal rates on social issues using a curated dataset of over 4,000 Wikipedia pages.
  • It employs regular audits—biweekly and weekly checks—to capture and analyze refusal patterns, reflecting evolving company policies and societal influences.
  • Results indicate significant differences across models, with GPT-4.1 at 3.9% and GPT-5 at 1.2% refusal rates, and highlight shifts tied to geopolitical events.

Longitudinal Monitoring of LLM Content Moderation of Social Issues

Introduction

The paper "Longitudinal Monitoring of LLM Content Moderation of Social Issues" introduces the AI Watchman system, a tool designed to audit and track the refusal rates of LLMs over time when prompted with content related to social issues. The primary focus is on measuring how these models, particularly OpenAI’s GPT-4.1, GPT-5, and DeepSeek, handle content moderation, specifically refusals, and how this reflects underlying company policies and potentially shapes public discourse.

AI Watchman System Overview

AI Watchman is structured to provide transparency into LLMs’ dynamic moderation practices. It operates by generating prompts from a curated dataset of social issues, engaging the model APIs, and systematically capturing instances where the models refuse to replicate given content, either wholly or partially. Figure 1

Figure 1: An overview of the AI Watchman\ system.

Key components of the system include:

  • Prompt Generation: Utilizing over 4,000 Wikipedia pages categorized into 52 social issue topics, the system formulates prompts to challenge the LLMs with requests to ā€œrepeat after me,ā€ assessing their refusal to engage with certain content.
  • Regular Monitoring and Data Capture: The system is programmed to perform audits on a biweekly basis for LLMs and weekly for OpenAI’s moderation endpoint, ensuring a consistent and comprehensive data collection.
  • Refusal Analysis: AI Watchman interprets refusals based on predefined patterns in response language or structured flags, providing a quantitative flagging rate across categories.
  • Public Visualization: It makes results publicly visible through a website with interactive graphs showcasing flagging trends over time. Figure 2

    Figure 2: The AI Watchman\ website as shown when it first loads.

Findings

Through AI Watchman, the researchers discovered notable differences in refusal rates and patterns between models. GPT-4.1 exhibited the highest refusal rate at 3.9%, whereas GPT-5, designed to mitigate explicit refusals, showed a significantly reduced rate of 1.2%. Notably, the moderation endpoint flagged content at a rate reflecting mostly violence-related concerns.

Content Moderation Patterns

  • OpenAI’s Models: GPT-4.1 tended to refuse content related to U.S. political figures, politics, and government, whereas GPT-5 shifted focus towards more nuanced decisions involving abortion-related topics recently.
  • DeepSeek: Demonstrated higher refusal rates for "Chinese Sensitive Topics," aligning with state policies, highlighting how geographical and political contexts influence AI moderation. Figure 3

    Figure 3: This bar chart illustrates the average refusal rate across models. Data included is from September 2025. Average rates at which the ``repeat after me'' query followed by the Social Issues Dataset Wikipedia content is refused varies from 1.2\% to 3.9\% depending on the model.

Temporal Changes and Implications

Substantial variance in moderation practices over time was evident, with notable shifts during specific geopolitical events:

  • Israel: GPT-4.1 refusals increased dramatically during the escalation of the Israel-Gaza conflict.
  • Abortion: A recent spike in GPT-5 refusals corresponds with legislative activities in Texas regarding abortion medication. Figure 4

Figure 4

Figure 4: Refusal rates for geopolitical content categories in GPT-4.1 over time. Israel-related content shows an increase from 20\% to 60\% refusal rates between August 18 and September 1, 2025, while other related topics remain relatively stable. Left: the Global Image category as shown in the AI Watchman\ overview. Right: the per-topic trend lines.

Discussion

LLMs, akin to search engines, are mediators of information, and their refusals reflect corporate values and societal norms. As LLMs’ role in content distribution expands, understanding their moderation mechanisms becomes crucial for evaluating their impact on public discourse. Despite a gradual shift towards "safe-completion" strategies, which may obscure refusals, AI Watchman provides a lens to scrutinize these systems transparently.

Conclusion

AI Watchman serves as a pivotal system for auditing LLMs, providing insights into their refusal patterns and offering a basis for public scrutiny. This investigation deepens the understanding of how AI systems mediate social issues, emphasizing the importance of longitudinal auditing to ensure accountability and transparency in AI-driven platforms.

The implications of these findings stretch beyond mere technical assessment into a broader societal context, investigating how automated systems influence public access to information and shape narratives around sensitive subjects. With the continuous integration of AI in everyday information-seeking behaviors, facilitating open discourse around their moderation practices is essential for equitable access to information.

Dice Question Streamline Icon: https://streamlinehq.com

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets

This paper has been mentioned in 1 tweet and received 0 likes.

Upgrade to Pro to view all of the tweets about this paper: