Beyond Hate Speech: NLP's Challenges and Opportunities in Uncovering Dehumanizing Language (2402.13818v1)
Abstract: Dehumanization, characterized as a subtle yet harmful manifestation of hate speech, involves denying individuals of their human qualities and often results in violence against marginalized groups. Despite significant progress in Natural Language Processing across various domains, its application in detecting dehumanizing language is limited, largely due to the scarcity of publicly available annotated data for this domain. This paper evaluates the performance of cutting-edge NLP models, including GPT-4, GPT-3.5, and LLAMA-2, in identifying dehumanizing language. Our findings reveal that while these models demonstrate potential, achieving a 70\% accuracy rate in distinguishing dehumanizing language from broader hate speech, they also display biases. They are over-sensitive in classifying other forms of hate speech as dehumanization for a specific subset of target groups, while more frequently failing to identify clear cases of dehumanization for other target groups. Moreover, leveraging one of the best-performing models, we automatically annotated a larger dataset for training more accessible models. However, our findings indicate that these models currently do not meet the high-quality data generation threshold necessary for this task.
- Open-source large language models outperform crowd workers and approach chatgpt in text-annotation tasks. arXiv preprint arXiv:2307.02179.
- Scaling instruction-finetuned language models. arXiv preprint arXiv:2210.11416.
- Race-based biases in judgments of social pain. Journal of Experimental Social Psychology, 88:103964.
- Is GPT-3 a good data annotator? In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 11173–11195, Toronto, Canada. Association for Computational Linguistics.
- Markus Eberts and Adrian Ulges. 2020. Span-based joint entity and relation extraction with transformer pre-training. In 24th European Conference on Artificial Intelligence - ECAI 2020, pages 2006–2013. IOS Press.
- Toward transformer-based nlp for extracting psychosocial indicators of moral disengagement. In Proceedings of the Annual Meeting of the Cognitive Science Society, volume 43.
- Chatgpt outperforms crowd-workers for text-annotation tasks. arXiv preprint arXiv:2303.15056.
- Liberals and conservatives rely on different sets of moral foundations. Journal of personality and social psychology, 96(5):1029.
- John Hagan and Wenona Rymond-Richmond. 2008. The collective dynamics of racial dehumanization and genocidal victimization in darfur. American Sociological Review, 73(6):875–902.
- Lasana T Harris and Susan T Fiske. 2015. Dehumanized perception. Zeitschrift für Psychologie.
- Nick Haslam. 2006. Dehumanization: An integrative review. Personality and social psychology review, 10(3):252–264.
- Subhuman, inhuman, and superhuman: Contrasting humans with nonhumans in three cultures. Social cognition, 26(2):248–258.
- Nick Haslam and Steve Loughnan. 2014. Dehumanization and infrahumanization. Annual review of psychology, 65:399–423.
- Annollm: Making large language models to be better crowdsourced annotators. arXiv preprint arXiv:2303.16854.
- Nour S. Kteily and Alexander P. Landry. 2022. Dehumanization: trends, insights, and challenges. Trends in Cognitive Sciences, 26(3):222–240.
- Can large language models aid in annotating speech emotional data? uncovering new frontiers. arXiv preprint arXiv:2307.06090.
- The emotional side of prejudice: The attribution of secondary emotions to ingroups and outgroups. Personality and social psychology review, 4(2):186–197.
- Roberta: A robustly optimized bert pretraining approach.
- A framework for the computational linguistic analysis of dehumanization. Frontiers in Artificial Intelligence, 3.
- Efficient estimation of word representations in vector space. arXiv preprint arXiv:1301.3781.
- Saif Mohammad. 2018. Obtaining reliable human ratings of valence, arousal, and dominance for 20,000 english words. In Proceedings of the 56th annual meeting of the association for computational linguistics (volume 1: Long papers), pages 174–184.
- Anthony Oberschall. 1997. Vojislav Seselj’s nationalist propaganda: contents, techniques, aims and impacts, 1990-1994. How mass media propaganda impacts on ordinary people’s acceptance and participation in collective violence, and how Seselj’s nationalist propaganda promoted and justified coercion and violence by the Serbs against non-Serbs.
- Differential association of uniquely and non uniquely human emotions with the ingroup and the outgroup. Group Processes & Intergroup Relations, 5(2):105–117.
- Connotation frames: A data-driven investigation. arXiv preprint arXiv:1506.02739.
- Connotation frames: A data-driven investigation. In Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 311–321, Berlin, Germany. Association for Computational Linguistics.
- Connotation frames of power and agency in modern films. In Proceedings of the 2017 conference on empirical methods in natural language processing, pages 2329–2334.
- From humans to machines: can chatgpt-like llms effectively replace human annotators in nlp tasks. In Workshop Proceedings of the 17th International AAAI Conference on Web and Social Media.
- Llama: Open and efficient foundation language models. arXiv preprint arXiv:2302.13971.
- Llama 2: Open foundation and fine-tuned chat models.
- Petter Törnberg. 2023. Chatgpt-4 outperforms experts and crowd workers in annotating political twitter messages with zero-shot learning. arXiv preprint arXiv:2304.06588.
- Learning from the worst: Dynamically generated datasets to improve online hate detection. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pages 1667–1682, Online. Association for Computational Linguistics.
- Huggingface’s transformers: State-of-the-art natural language processing.
- Can large language models transform computational social science? arXiv preprint arXiv:2305.03514.