Supporting Qualitative Analysis with Large Language Models: Combining Codebook with GPT-3 for Deductive Coding

被引:53
|
作者
Xiao, Ziang [1 ]
Yuan, Xingdi [1 ]
Liao, Q. Vera [1 ]
Abdelghani, Rania [2 ]
Oudeyer, Pierre-Yves [2 ]
机构
[1] Microsoft Res, Montreal, PQ, Canada
[2] INRIA, Paris, France
关键词
Qualitative Analysis; Deductive Coding; Large Language Model; GPT-3;
D O I
10.1145/3581754.3584136
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Qualitative analysis of textual contents unpacks rich and valuable information by assigning labels to the data. However, this process is often labor-intensive, particularly when working with large datasets. While recent AI-based tools demonstrate utility, researchers may not have readily available AI resources and expertise, let alone be challenged by the limited generalizability of those task-specific models. In this study, we explored the use of large language models (LLMs) in supporting deductive coding, a major category of qualitative analysis where researchers use pre-determined code-books to label the data into a fixed set of codes. Instead of training task-specific models, a pre-trained LLM could be used directly for various tasks without fine-tuning through prompt learning. Using a curiosity-driven questions coding task as a case study, we found, by combining GPT-3 with expert-drafted codebooks, our proposed approach achieved fair to substantial agreements with expert-coded results. We lay out challenges and opportunities in using LLMs to support qualitative coding and beyond.
引用
收藏
页码:75 / 78
页数:4
相关论文
共 48 条