Using Large Pretrained Language Models for Answering User Queries from Product Specifications

被引:0
|
作者
Roy, Kalyani [1 ]
Shah, Smit [1 ]
Pai, Nithish [2 ]
Ramtej, Jaidam [2 ]
Nadkarn, Prajit Prashant [2 ]
Banerjee, Jyotirmoy [2 ]
Goyal, Pawan [1 ]
Kumar, Surender [2 ]
机构
[1] Indian Inst Technol Kharagpur, Kharagpur, W Bengal, India
[2] Flipkart, Bengaluru, India
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
While buying a product from the c-eommcree websites, customers generally have a plethora of questions. From the perspective of both the e-commerce service provider as well as the customers, there must be an effective question answering system to provide immediate answers to the user queries. While certain questions can only be answered after using the product, there are many questions which can be answered from the product specification itself. Our work takes a first step in this direction by finding out the relevant product specifications, that can help answering the user questions. We propose an approach to automatically create a training dataset for this problem. We utilize recently proposed XLNet and BERT architectures for this problem and find that they provide much better performance than the Siamese model, previously applied for this problem (Lai et al., 2018). Our model gives a good performance even when trained on one vertical and tested across different verticals.
引用
收藏
页码:35 / 39
页数:5
相关论文
共 50 条
  • [1] Large Product Key Memory for Pretrained Language Models
    Kim, Gyuwan
    Jung, Tae-Hwan
    [J]. FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2020, 2020, : 4060 - 4069
  • [2] Evolving Landscape of Large Language Models: An Evaluation of ChatGPT and Bard in Answering Patient Queries on Colonoscopy
    Tariq, Raseen
    Malik, Sheza
    Khanna, Sahil
    [J]. GASTROENTEROLOGY, 2024, 166 (01) : 220 - 221
  • [3] Leveraging Text-to-Text Pretrained Language Models for Question Answering in Chemistry
    Tran, Dan
    Pascazio, Laura
    Akroyd, Jethro
    Mosbach, Sebastian
    Kraft, Markus
    [J]. ACS OMEGA, 2024, 9 (12): : 13883 - 13896
  • [4] PROSPER: Extracting Protocol Specifications Using Large Language Models
    Sharma, Prakhar
    Yegneswaran, Vinod
    [J]. PROCEEDINGS OF THE 22ND ACM WORKSHOP ON HOT TOPICS IN NETWORKS, HOTNETS 2023, 2023, : 41 - 47
  • [5] Constructing Taxonomies from Pretrained Language Models
    Chen, Catherine
    Lin, Kevin
    Klein, Dan
    [J]. 2021 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL-HLT 2021), 2021, : 4687 - 4700
  • [6] Application of Pretrained Large Language Models in Embodied Artificial Intelligence
    A. K. Kovalev
    A. I. Panov
    [J]. Doklady Mathematics, 2022, 106 : S85 - S90
  • [7] Generalized Planning in PDDL Domains with Pretrained Large Language Models
    Silver, Tom
    Dan, Soham
    Srinivas, Kavitha
    Tenenbaum, Joshua B.
    Kaelbling, Leslie
    Katz, Michael
    [J]. THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 18, 2024, : 20256 - 20264
  • [8] Intent-based Product Collections for E-commerce using Pretrained Language Models
    Kim, Hiun
    Jeong, Jisu
    Kim, Kyung-Min
    Lee, Dongjun
    Lee, Hyun Dong
    Seo, Dongpil
    Han, Jeeseung
    Park, Dong Wook
    Heo, Ji Ae
    Kim, Rak Yeong
    [J]. 21ST IEEE INTERNATIONAL CONFERENCE ON DATA MINING WORKSHOPS ICDMW 2021, 2021, : 228 - 237
  • [9] Generating Specifications from Requirements Documents for Smart Devices Using Large Language Models (LLMs)
    Lutze, Rainer
    Waldhoer, Klemens
    [J]. HUMAN-COMPUTER INTERACTION, PT I, HCI 2024, 2024, 14684 : 94 - 108
  • [10] Application of Pretrained Large Language Models in Embodied Artificial Intelligence
    Kovalev, A. K.
    Panov, A. I.
    [J]. DOKLADY MATHEMATICS, 2022, 106 (SUPPL 1) : S85 - S90