Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
194 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
46 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

SemEval-2020 Task 11: Detection of Propaganda Techniques in News Articles (2009.02696v1)

Published 6 Sep 2020 in cs.CL and cs.CY

Abstract: We present the results and the main findings of SemEval-2020 Task 11 on Detection of Propaganda Techniques in News Articles. The task featured two subtasks. Subtask SI is about Span Identification: given a plain-text document, spot the specific text fragments containing propaganda. Subtask TC is about Technique Classification: given a specific text fragment, in the context of a full document, determine the propaganda technique it uses, choosing from an inventory of 14 possible propaganda techniques. The task attracted a large number of participants: 250 teams signed up to participate and 44 made a submission on the test set. In this paper, we present the task, analyze the results, and discuss the system submissions and the methods they used. For both subtasks, the best systems used pre-trained Transformers and ensembles.

Citations (208)

Summary

  • The paper introduces Transformer and ensemble-based systems that achieved an F1 score of 51.55 for propaganda span identification.
  • It leverages auxiliary tasks and semantic features to classify 14 propaganda techniques, reaching an F1 score of 62.07 in technique classification.
  • The study advances automatic media literacy by offering scalable strategies for detecting disinformation in news articles.

Detection of Propaganda Techniques in News Articles: A Summary of SemEval-2020 Task 11

The paper discusses the SemEval-2020 Task 11, which focused on the detection of propaganda techniques in news articles. This task was split into two subtasks: Span Identification (SI) and Technique Classification (TC). Subtask SI required identifying specific text fragments within a document where propaganda techniques are present. Subtask TC involved classifying these identified fragments into one of 14 predefined propaganda techniques.

Participants and Methodologies

The task attracted significant interest, with 250 registered teams and 44 submissions for the test phase. The most successful systems in both subtasks incorporated pre-trained Transformers such as BERT and RoBERTa and deployed ensemble learning strategies.

  1. Span Identification (SI): The leading system utilized a sophisticated architecture combining pre-trained models (BERT, RoBERTa) with additional layers like LSTM and CRF for sequence labeling. Auxiliary tasks were introduced to aid the primary task of span identification, enhancing performance. Many teams enhanced their models with domain-adapted LLMs and generated "silver" data to enlarge the training set effectively.
  2. Technique Classification (TC): Participants employed similar Transformer-based models, often combined with sentence embeddings and contextual features surrounding the propaganda span. Several teams achieved improvements by integrating auxiliary semantic features and engineered features such as sentiment and entity-related information.

Results

  • Span Identification: The top-performing system achieved an F1 score of 51.55 on the test set, demonstrating moderate success in identifying propaganda spans. While there was a notable variance in performance across systems, Transformer-based approaches outperformed others, confirming their efficacy in token-level sequence tasks.
  • Technique Classification: The highest rank in technique classification achieved an F1 score of 62.07. Key challenges remained in precisely distinguishing between the 14 techniques, some of which were less frequent and harder to detect. This task's complexity arose partly due to the infrequent and overlapping nature of certain propaganda techniques.

Implications and Future Directions

The task highlighted the potential of using advanced NLP models to perform fine-grained content analysis and detect manipulative language in news media. The insights gained from this paper provide valuable guidance for the development of real-world applications targeting media literacy and automatic disinformation detection.

Future work can explore the integration of multimodal data to enhance the context understanding and analysis of propaganda. Additionally, expanding datasets to include a broader range and volume of annotated data will be crucial for training more generalized models capable of operating across diverse languages and media formats. Further research may also delve into real-time system implementations and evaluate their efficacy in different media and geopolitical contexts.

In conclusion, while the task laid foundational work in identifying and classifying propaganda in news articles, ongoing advancements in AI and computational linguistics will continue to proliferate robust solutions in mitigating the influence of propaganda and enhancing the veracity of information disseminated to the public.