A Quantum Annealing Instance Selection Approach for Efficient and Effective Transformer Fine-Tuning

被引:0
|
作者
Pasin, Andrea [1 ]
Cunha, Washington [2 ]
Goncalves, Marcos Andre [2 ]
Ferro, Nicola [1 ]
机构
[1] Univ Padua, Padua, Italy
[2] Univ Fed Minas Gerais, Belo Horizonte, MG, Brazil
基金
巴西圣保罗研究基金会;
关键词
Instance Selection; Quantum Computing; Text Classification;
D O I
10.1145/3664190.3672515
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Deep Learning approaches have become pervasive in recent years due to their ability to solve complex tasks. However, these models need huge datasets for proper training and good generalization. This translates into high training and fine-tuning time, even several days for the most complex models and large datasets. In this work, we present a novel quantum Instance Selection (IS) approach that allows to significantly reduce the size of the training datasets (by up to 28%) while maintaining the model's effectiveness, thus promoting (training) speedups and scalability. Our solution is innovative in the sense that it exploits a different computing paradigm - Quantum Annealing (QA) - a specific Quantum Computing paradigm that can be used to tackle optimization problems. To the best of our knowledge, there have been no prior attempts to tackle the IS problem using QA. Furthermore, we propose a new Quadratic Unconstrained Binary Optimization formulation specific for the IS problem, which is a contribution in itself. Through an extensive set of experiments with several Text Classification benchmarks, we empirically demonstrate our quantum solution's feasibility and competitiveness with the current state-of-the-art IS solutions.
引用
收藏
页码:205 / 214
页数:10
相关论文
共 50 条
  • [31] Fine-tuning and vacuum stability in the Wilsonian effective action
    Krajewski, Tomasz
    Lalak, Zygmunt
    PHYSICAL REVIEW D, 2015, 92 (07)
  • [32] Fine-Tuning of the Supratip in Rhinoplasty: An External Approach
    Arda Kucukguven
    Ersoy Konas
    Aesthetic Plastic Surgery, 2022, 46 : 2938 - 2946
  • [33] DeepHash for Image Instance Retrieval: Getting Regularization, Depth and Fine-Tuning Right
    Lin, Jie
    Morere, Olivier
    Veillard, Antoine
    Duan, Ling-Yu
    Goh, Hanlin
    Chandrasekhar, Vijay
    PROCEEDINGS OF THE 2017 ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA RETRIEVAL (ICMR'17), 2017, : 138 - 146
  • [34] How fine can fine-tuning be? Learning efficient language models
    Radiya-Dixit, Evani
    Wang, Xin
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 108, 2020, 108 : 2435 - 2442
  • [35] Augmented Neural Fine-Tuning for Efficient Backdoor Purification
    Karim, Nazmul
    Al Arafat, Abdullah
    Khalid, Umar
    Guo, Zhishan
    Rahnavard, Nazanin
    COMPUTER VISION - ECCV 2024, PT LXXX, 2025, 15138 : 401 - 418
  • [36] Fine-tuning pretrained transformer encoders for sequence-to-sequence learning
    Bao, Hangbo
    Dong, Li
    Wang, Wenhui
    Yang, Nan
    Piao, Songhao
    Wei, Furu
    INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2024, 15 (05) : 1711 - 1728
  • [37] Domain Adaptation for Efficiently Fine-tuning Vision Transformer with Encrypted Images
    Nagamori, Teru
    Shiota, Sayaka
    Kiya, Hitoshi
    2023 ASIA PACIFIC SIGNAL AND INFORMATION PROCESSING ASSOCIATION ANNUAL SUMMIT AND CONFERENCE, APSIPA ASC, 2023, : 1677 - 1683
  • [38] Fine-tuning Transformer-based MT Using Syntactic Guides
    Makwisai, Jinnawat
    Boonkwan, Prachya
    Usanavasin, Sasiporn
    Kaothanthong, Natsuda
    Okumura, Manabu
    2023 18TH INTERNATIONAL JOINT SYMPOSIUM ON ARTIFICIAL INTELLIGENCE AND NATURAL LANGUAGE PROCESSING, ISAI-NLP, 2023,
  • [39] Fine-tuning pretrained transformer encoders for sequence-to-sequence learning
    Hangbo Bao
    Li Dong
    Wenhui Wang
    Nan Yang
    Songhao Piao
    Furu Wei
    International Journal of Machine Learning and Cybernetics, 2024, 15 : 1711 - 1728
  • [40] Evaluation of Dataset Selection for Pre-Training and Fine-Tuning Transformer Language Models for Clinical Question Answering
    Soni, Sarvesh
    Roberts, Kirk
    PROCEEDINGS OF THE 12TH INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION (LREC 2020), 2020, : 5532 - 5538