Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Large Language Models and Thematic Analysis: Human-AI Synergy in Researching Hate Speech on Social Media (2408.05126v1)

Published 9 Aug 2024 in cs.HC, cs.CL, and cs.SI

Abstract: In the dynamic field of AI, the development and application of LLMs for text analysis are of significant academic interest. Despite the promising capabilities of various LLMs in conducting qualitative analysis, their use in the humanities and social sciences has not been thoroughly examined. This article contributes to the emerging literature on LLMs in qualitative analysis by documenting an experimental study involving GPT-4. The study focuses on performing thematic analysis (TA) using a YouTube dataset derived from an EU-funded project, which was previously analyzed by other researchers. This dataset is about the representation of Roma migrants in Sweden during 2016, a period marked by the aftermath of the 2015 refugee crisis and preceding the Swedish national elections in 2017. Our study seeks to understand the potential of combining human intelligence with AI's scalability and efficiency, examining the advantages and limitations of employing LLMs in qualitative research within the humanities and social sciences. Additionally, we discuss future directions for applying LLMs in these fields.

A Detailed Examination of LLMs and Thematic Analysis in Social Media Hate Speech Research

In the landscape of AI, the application of LLMs in text analysis has garnered substantial attention. The paper "LLMs and Thematic Analysis: Human-AI Synergy in Researching Hate Speech on Social Media" by Breazu et al. explores the experimental integration of GPT-4 to conduct thematic analysis (TA) on a dataset comprising YouTube comments concerning Roma migrants in Sweden. This paper aims to illuminate the synergy between human intelligence and AI, assessing the efficacy, limitations, and future applications of LLMs in qualitative research within the humanities and social sciences.

Experimental Setup and Methodology

The experiment utilized OpenAI's state-of-the-art LLM, GPT-4, to perform TA on a YouTube dataset from an EU-funded project previously analyzed by researchers. The dataset focused on comments about Roma migrants in Sweden during 2016, a period influenced by the 2015 refugee crisis and preceding the 2017 Swedish national elections. The paper employed both inductive and deductive approaches to TA, defined by Braun and Clarke (2006), to examine GPT-4's capability in identifying themes independently (inductive) and within a structured framework (deductive).

Initially, GPT-4 was assigned the task of categorizing the comments inductively. The model processed the comments in batches and generated themes which were later compared with the categories derived by a human researcher. Subsequently, predefined categories were fed into GPT-4 to perform a deductive analysis, assigning each comment to an established category. Throughout the process, the categorization quality was evaluated by four domain experts in thematic analysis.

Findings

Initial Classification and Theme Extraction:

GPT-4's inductive analysis yielded 152 initial themes, which were distilled into five main categories after redundancy elimination and theme refinement: Ethnic Misunderstanding and Identity Confusion, Stereotyping and Social Prejudice, Economic Concerns and Welfare Debates, Cultural Clash and Integration Challenges, and Polarization of Public Opinion. This initial classification was aligned well with the human researcher's themes, indicating GPT-4's potential in identifying broad themes within qualitative datasets.

Human-Supported Categorization:

When tasked to categorize the data using human-defined categories, GPT-4's analysis revealed the necessity of context-specific training. The human researcher’s categories, such as Populism, Nativism, Extreme Hate Speech, Cultural Racism, (Non)Belonging, and Perceptions and Stereotypes, provided a more detailed and nuanced understanding of the comments. The comparison highlighted differences in the depth and specificity brought by human expertise, which GPT-4’s broader, neutral approach lacked.

Implications and Future Directions:

The paper underscores the necessity for human oversight in qualitative thematic analysis, even as LLMs like GPT-4 show promising abilities in processing and categorizing large datasets. By incorporating detailed context learning and theory-driven prompts, researchers can better align LLM outputs with specific analytical frameworks and socio-political contexts. This synergy can significantly enhance the accuracy and comprehensiveness of qualitative research.

Ethical Considerations and Limitations:

Acknowledging the ethical considerations and limitations associated with using LLMs, particularly in analyzing sensitive data such as hate speech, is crucial. The authors highlighted potential issues like model drift, hallucinations, and cautious handling of sensitive data to ensure ethical standards and reliability.

Conclusion

The integration of LLMs in qualitative research offers unprecedented opportunities for enhancing the scope and efficiency of thematic analysis. However, careful management is essential to ensure that AI complements rather than replaces human expertise. This paper by Breazu et al. exemplifies the potential and limitations of LLMs in qualitative analysis, advocating for a synergistic approach that harnesses the strengths of both AI and human intelligence.

Future developments should focus on refining LLM capabilities through context-specific training, incorporating detailed theoretical frameworks into prompts, and addressing the ethical and practical challenges of using AI in social science research. By doing so, researchers can leverage the scalability and efficiency of LLMs while maintaining the critical insights derived from human expertise.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Petre Breazu (1 paper)
  2. Miriam Schirmer (6 papers)
  3. Songbo Hu (9 papers)
  4. Napoleon Katsos (1 paper)
X Twitter Logo Streamline Icon: https://streamlinehq.com