Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
38 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Supporting Qualitative Analysis with Large Language Models: Combining Codebook with GPT-3 for Deductive Coding (2304.10548v1)

Published 17 Apr 2023 in cs.CL, cs.AI, and cs.HC

Abstract: Qualitative analysis of textual contents unpacks rich and valuable information by assigning labels to the data. However, this process is often labor-intensive, particularly when working with large datasets. While recent AI-based tools demonstrate utility, researchers may not have readily available AI resources and expertise, let alone be challenged by the limited generalizability of those task-specific models. In this study, we explored the use of LLMs in supporting deductive coding, a major category of qualitative analysis where researchers use pre-determined codebooks to label the data into a fixed set of codes. Instead of training task-specific models, a pre-trained LLM could be used directly for various tasks without fine-tuning through prompt learning. Using a curiosity-driven questions coding task as a case study, we found, by combining GPT-3 with expert-drafted codebooks, our proposed approach achieved fair to substantial agreements with expert-coded results. We lay out challenges and opportunities in using LLMs to support qualitative coding and beyond.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Ziang Xiao (25 papers)
  2. Xingdi Yuan (46 papers)
  3. Q. Vera Liao (49 papers)
  4. Rania Abdelghani (7 papers)
  5. Pierre-Yves Oudeyer (95 papers)
Citations (96)