Supporting Qualitative Analysis with Large Language Models: Combining Codebook with GPT-3 for Deductive Coding

被引:53
|
作者
Xiao, Ziang [1 ]
Yuan, Xingdi [1 ]
Liao, Q. Vera [1 ]
Abdelghani, Rania [2 ]
Oudeyer, Pierre-Yves [2 ]
机构
[1] Microsoft Res, Montreal, PQ, Canada
[2] INRIA, Paris, France
关键词
Qualitative Analysis; Deductive Coding; Large Language Model; GPT-3;
D O I
10.1145/3581754.3584136
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Qualitative analysis of textual contents unpacks rich and valuable information by assigning labels to the data. However, this process is often labor-intensive, particularly when working with large datasets. While recent AI-based tools demonstrate utility, researchers may not have readily available AI resources and expertise, let alone be challenged by the limited generalizability of those task-specific models. In this study, we explored the use of large language models (LLMs) in supporting deductive coding, a major category of qualitative analysis where researchers use pre-determined code-books to label the data into a fixed set of codes. Instead of training task-specific models, a pre-trained LLM could be used directly for various tasks without fine-tuning through prompt learning. Using a curiosity-driven questions coding task as a case study, we found, by combining GPT-3 with expert-drafted codebooks, our proposed approach achieved fair to substantial agreements with expert-coded results. We lay out challenges and opportunities in using LLMs to support qualitative coding and beyond.
引用
收藏
页码:75 / 78
页数:4
相关论文
共 48 条
  • [41] ADVANCING SYSTEMATIC LITERATURE REVIEWS: A COMPARATIVE ANALYSIS OF LARGE LANGUAGE MODELS (CLAUDE SONNET 3.5, GEMINI FLASH 1.5, AND GPT-4) IN THE AUTOMATION ERA OF GENERATIVE AI
    Rai, P.
    Pandey, S.
    Attri, S.
    Singh, B.
    Kaur, R.
    VALUE IN HEALTH, 2024, 27 (12)
  • [42] Inductive Thematic Analysis of Healthcare Qualitative Interviews Using Open-Source Large Language Models: How Does it Compare to Traditional Methods?
    Mathis, Walter S.
    Zhao, Sophia
    Pratt, Nicholas
    Weleff, Jeremy
    De Paoli, Stefano
    SSRN,
  • [43] Inductive thematic analysis of healthcare qualitative interviews using open-source large language models: How does it compare to traditional methods?
    Mathis, Walter S.
    Zhao, Sophia
    Pratt, Nicholas
    Weleff, Jeremy
    De Paoli, Stefano
    COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE, 2024, 255
  • [44] Assessing Economic Viability: A Comparative Analysis of Total Cost of Ownership for Domain-Adapted Large Language Models versus State-of-the-art Counterparts in Chip Design Coding Assistance
    Sharma, Amit
    Ene, Teodor-Dumitru
    Kunal, Kishor
    Liu, Mingjie
    Hasan, Zafar
    Ren, Haoxing
    2024 IEEE LLM AIDED DESIGN WORKSHOP, LAD 2024, 2024,
  • [45] Large Language Models in an App: Conducting a Qualitative Synthetic Data Analysis of How Snapchat's "My AI" Responds to Questions About Sexual Consent, Sexual Refusals, Sexual Assault, and Sexting
    Marcantonio, Tiffany L.
    Avery, Gracie
    Thrash, Anna
    Leone, Ruschelle M.
    JOURNAL OF SEX RESEARCH, 2024,
  • [46] Benchmarking the performance of large language models in uveitis: a comparative analysis of ChatGPT-3.5, ChatGPT-4.0, Google Gemini, and Anthropic Claude3
    Zhao, Fang-Fang
    He, Han-Jie
    Liang, Jia-Jian
    Cen, Jingyun
    Wang, Yun
    Lin, Hongjie
    Chen, Feifei
    Li, Tai-Ping
    Yang, Jian-Feng
    Chen, Lan
    Cen, Ling-Ping
    EYE, 2024,
  • [47] Comment on: "Benchmarking the performance of large language models in uveitis: a comparative analysis of ChatGPT-3.5, ChatGPT-4.0, Google Gemini, and Anthropic Claude3"
    Luo, Xiao
    Tang, Cheng
    Chen, Jin-Jin
    Yuan, Jin
    Huang, Jin-Jin
    Yan, Tao
    EYE, 2025, : 1432 - 1432
  • [48] Reply to 'Comment on: Benchmarking the performance of large language models in uveitis: a comparative analysis of ChatGPT-3.5, ChatGPT-4.0, Google Gemini, and Anthropic Claude3'
    Zhao, Fang-Fang
    He, Han-Jie
    Liang, Jia-Jian
    Cen, Ling-Ping
    EYE, 2025, : 1433 - 1433