Learning a Better Negative Sampling Policy with Deep Neural Networks for Search

被引:6
|
作者
Cohen, Daniel [1 ]
Jordan, Scott M. [2 ]
Croft, W. Bruce [1 ]
机构
[1] Univ Massachusetts Amherst, Ctr Intelligent Informat Retrieval, Amherst, MA 01003 USA
[2] Univ Massachusetts Amherst, Autonomous Learning Lab, Amherst, MA 01003 USA
关键词
D O I
10.1145/3341981.3344220
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
In information retrieval, sampling methods used to select documents for neural models must often deal with large class imbalances during training. This issue necessitates careful selection of negative instances when training neural models to avoid the risk of overfitting. For most work, heuristic sampling approaches, or policies, are created based off of domain experts, such as choosing samples with high BM25 scores or a random process over candidate documents. However, these sampling approaches are done with the test distribution in mind. In this paper, we demonstrate that the method chosen to sample negative documents during training plays a critical role in both the stability of training, as well as overall performance. Furthermore, we establish that using reinforcement learning to optimize a policy over a set of sampling functions can significantly improve performance over standard training practices with respect to IR metrics and is robust to hyperparameters and random seeds.
引用
收藏
页码:19 / 26
页数:8
相关论文
共 50 条
  • [1] Learning deep morphological networks with neural architecture search
    Hu, Yufei
    Belkhir, Nacim
    Angulo, Jesus
    Yao, Angela
    Franchi, Gianni
    PATTERN RECOGNITION, 2022, 131
  • [2] Sampling weights of deep neural networks
    Bolager, Erik Lien
    Burak, Iryna
    Datar, Chinmay
    Sun, Qing
    Dietrich, Felix
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [3] Mixed Negative Sampling for Learning Two-tower Neural Networks in Recommendations
    Yang, Ji
    Yi, Xinyang
    Cheng, Derek Zhiyuan
    Hong, Lichan
    Li, Yang
    Wang, Simon Xiaoming
    Xu, Taibai
    Chi, Ed H.
    WWW'20: COMPANION PROCEEDINGS OF THE WEB CONFERENCE 2020, 2020, : 441 - 447
  • [4] Deep or Wide? Learning Policy and Value Neural Networks for Combinatorial Games
    Edelkamp, Stefan
    COMPUTER GAMES: 5TH WORKSHOP ON COMPUTER GAMES, CGW 2016, AND 5TH WORKSHOP ON GENERAL INTELLIGENCE IN GAME-PLAYING AGENTS, GIGA 2016, HELD IN CONJUNCTION WITH THE 25TH INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2016, NEW YORK, USA, JULY 9-10, 2016, 2017, 705 : 19 - 33
  • [5] Search for deep graph neural networks
    Feng, Guosheng
    Wang, Hongzhi
    Wang, Chunnan
    INFORMATION SCIENCES, 2023, 649
  • [6] Rolling the dice for better deep learning performance: A study of randomness techniques in deep neural networks
    Altarabichi, Mohammed Ghaith
    Nowaczyk, Slawomir
    Pashami, Sepideh
    Mashhadi, Peyman Sheikholharam
    Handl, Julia
    INFORMATION SCIENCES, 2024, 667
  • [7] Towards Better Generalization of Deep Neural Networks via Non-Typicality Sampling Scheme
    Peng, Xinyu
    Wang, Fei-Yue
    Li, Li
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, 34 (10) : 7910 - 7920
  • [8] TOWARDS BETTER UNCERTAINTY SAMPLING: ACTIVE LEARNING WITH MULTIPLE VIEWS FOR DEEP CONVOLUTIONAL NEURAL NETWORK
    He, Tao
    Jin, Xiaoming
    Ding, Guiguang
    Yi, Lan
    Yan, Chenggang
    2019 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO (ICME), 2019, : 1360 - 1365
  • [9] Deep neural networks understand investors better
    Mahmoudi, Nader
    Docherty, Paul
    Moscato, Pablo
    DECISION SUPPORT SYSTEMS, 2018, 112 : 23 - 34
  • [10] Deep Embedding Learning with Discriminative Sampling Policy
    Duan, Yueqi
    Chen, Lei
    Lu, Jiwen
    Zhou, Jie
    2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 4959 - 4968