Computational Thematic Analysis of Poetry via Bimodal Large Language Models

被引:1
|
作者
Choi K. [1 ]
机构
[1] Indiana University, United States
关键词
auxiliary data; computational poetry analysis; context-dependent language model; Digital libraries; multilabel classification;
D O I
10.1002/pra2.812
中图分类号
学科分类号
摘要
This article proposes a multilabel poem topic classification algorithm utilizing large language models and auxiliary data to address the lack of diverse metadata in digital poetry libraries. The study examines the potential of context-dependent language models, specifically bidirectional encoder representations from transformers (BERT), for understanding poetic words and utilizing auxiliary data, such as author's notes, in supplementing poetry text. The experimental results demonstrate that the BERT-based model outperforms the traditional support vector machine-based model across all input types and datasets. We also show that incorporating notes as an additional input improves the performance of the poem-only model. Overall, the study suggests pretrained context-dependent language models and auxiliary data have potential to enhance the accessibility of various poems within collections. This research can eventually assist in promoting the discovery of underrepresented poems in digital libraries, even if they lack associated metadata, thus enhancing the understanding and appreciation of the literary form. Annual Meeting of the Association for Information Science & Technology | Oct. 27 – 31, 2023 | London, United Kingdom. Author(s) retain copyright, but ASIS&T receives an exclusive publication license.
引用
收藏
页码:538 / 542
页数:4
相关论文
共 50 条
  • [21] D LLM-in-the-loop: Leveraging Large Language Model for Thematic Analysis
    Dai, Shih-Chieh
    Xiong, Aiping
    Ku, Lun-Wei
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EMNLP 2023), 2023, : 9993 - 10001
  • [22] Large Language Models Are Zero-Shot Fuzzers: Fuzzing Deep-Learning Libraries via Large Language Models
    Deng, Yinlin
    Xia, Chunqiu Steven
    Peng, Haoran
    Yang, Chenyuan
    Zhan, Lingming
    PROCEEDINGS OF THE 32ND ACM SIGSOFT INTERNATIONAL SYMPOSIUM ON SOFTWARE TESTING AND ANALYSIS, ISSTA 2023, 2023, : 423 - 435
  • [23] Exploring Automated Assertion Generation via Large Language Models
    Zhang, Quanjun
    Sun, Weifeng
    Fang, Chunrong
    Yu, Bowen
    Li, Hongyan
    Yan, Meng
    Zhou, Jianyi
    Chen, Zhenyu
    ACM TRANSACTIONS ON SOFTWARE ENGINEERING AND METHODOLOGY, 2025, 34 (03)
  • [24] Guiding Large Language Models via Directional Stimulus Prompting
    Li, Zekun
    Peng, Baolin
    He, Pengcheng
    Galley, Michel
    Gao, Jianfeng
    Yan, Xifeng
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [25] Aspects of Uncertainty Analysis for Large Nonlinear Computational Models
    Worden, K.
    Becker, W. E.
    Battipede, M.
    Surace, C.
    ADVANCES IN EXPERIMENTAL MECHANICS VII, 2010, 24-25 : 25 - +
  • [26] Computational Models of Language Acquisition
    Wintner, Shuly
    COMPUTATIONAL LINGUISTICS AND INTELLIGENT TEXT PROCESSING, 2010, 6008 : 86 - 99
  • [27] COMPUTATIONAL MODELS OF LANGUAGE PROCESSING
    STABLER, EP
    BEHAVIORAL AND BRAIN SCIENCES, 1986, 9 (03) : 550 - 551
  • [28] Computational complexity in language models
    Savitch, W
    ISSUES IN MATHEMATICAL LINGUISTICS, 1999, 47 : 183 - 191
  • [29] Using Large Language Models to Improve Sentiment Analysis in Latvian Language
    Purvins, Pauls
    Urtans, Evalds
    Caune, Vairis
    BALTIC JOURNAL OF MODERN COMPUTING, 2024, 12 (02): : 165 - 175
  • [30] On Teaching Novices Computational Thinking by Utilizing Large Language Models Within Assessments
    Hassan, Mohammed
    Chen, Yuxuan
    Denny, Paul
    Zilles, Craig
    PROCEEDINGS OF THE 56TH ACM TECHNICAL SYMPOSIUM ON COMPUTER SCIENCE EDUCATION, SIGCSE TS 2025, VOL 2, 2025, : 471 - 477