Korean Offensive Language Dataset: A Comprehensive Framework for Multi-Dimensional Offensive Language Analysis
The paper "KOLD: Korean Offensive Language Dataset" introduces the Korean Offensive Language Dataset (KOLD), which represents a substantial advancement in language-specific offensive language detection models through the creation of an annotated dataset tailored to Korean sociocultural contexts. The focus of this paper is to address the inadequacies of English-centric models and datasets when applied to Korean and potentially other non-English languages.
Dataset Development and Annotation
KOLD encompasses 40,429 comments sourced from NAVER news articles and YouTube videos, annotated with a hierarchical taxonomy that includes offensiveness, target types, target groups, and contextual information. The dataset provides annotations of offensive spans, target spans, and targeted groups tailored to Korean culture, addressing previous limitations in Korean datasets which lacked hierarchical and contextual markers.
Hierarchical Annotation Framework
The dataset is structured into three levels of annotation:
- Offensive Language Detection (Level A): This involves identifying whether a comment is offensive and pinpointing the offensive span within the text.
- Target Type Categorization (Level B): This level distinguishes whether the offensive comment is untargeted, aimed at an individual, a group, or other entities such as organizations.
- Target Group Identification (Level C): For group-targeted offenses, this identifies specific groups such as gender, ethnicity, political affiliation, or cultural identity.
Methodology and Evaluation
The paper employs various machine learning models, primarily Korean BERT and RoBERTa models, to evaluate KOLD's efficacy in offensiveness detection and categorization tasks. Noteworthy numerical results demonstrate improvements in model performance with title context information, significantly enhancing target group classification accuracy by 13.1% — a figure highlighting the impact of incorporating contextual data. The multi-task models outperform single-task models in span prediction, underscoring the value of joint learning in improving interpretability and span accuracy.
Cultural and Linguistic Insights
A striking observation from the dataset analysis is the cultural divergence in target group prevalence compared to English datasets. For instance, groups like Korean-Chinese and Feminists are more prominent targets in the Korean context, which are less frequently addressed in English datasets. This highlights linguistic nuances and sociopolitical differences between the communities, necessitating language-specific datasets for effective offensive language detection.
Implications and Future Work
The development of KOLD lays foundational groundwork for both theoretical and practical applications in NLP for non-English languages. On a theoretical level, it emphasizes the importance of culturally and linguistically specific approaches to hate speech detection. Practically, it provides a robust dataset for the development of models capable of adapting to complex cultural contexts, offering key insights for similar endeavors in other languages and regions.
Furthermore, the paper suggests the need for continuous updates to offensive language datasets to reflect evolving social and political climates, such as emergent hate speech patterns during crises like COVID-19. It hints at future work that could explore machine-generated datasets to balance scale and accuracy.
In summary, KOLD represents a significant step towards understanding and mitigating online offensive speech through a culturally informed lens. By publicly releasing this dataset, the authors invite further research aimed at enhancing offensive language detection systems globally, while also acknowledging ethical considerations related to the sensitive nature of the content.