Learning a Better Negative Sampling Policy with Deep Neural Networks for Search

被引:6
|
作者
Cohen, Daniel [1 ]
Jordan, Scott M. [2 ]
Croft, W. Bruce [1 ]
机构
[1] Univ Massachusetts Amherst, Ctr Intelligent Informat Retrieval, Amherst, MA 01003 USA
[2] Univ Massachusetts Amherst, Autonomous Learning Lab, Amherst, MA 01003 USA
关键词
D O I
10.1145/3341981.3344220
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
In information retrieval, sampling methods used to select documents for neural models must often deal with large class imbalances during training. This issue necessitates careful selection of negative instances when training neural models to avoid the risk of overfitting. For most work, heuristic sampling approaches, or policies, are created based off of domain experts, such as choosing samples with high BM25 scores or a random process over candidate documents. However, these sampling approaches are done with the test distribution in mind. In this paper, we demonstrate that the method chosen to sample negative documents during training plays a critical role in both the stability of training, as well as overall performance. Furthermore, we establish that using reinforcement learning to optimize a policy over a set of sampling functions can significantly improve performance over standard training practices with respect to IR metrics and is robust to hyperparameters and random seeds.
引用
收藏
页码:19 / 26
页数:8
相关论文
共 50 条
  • [41] Better schedules for low precision training of deep neural networks
    Wolfe, Cameron R.
    Kyrillidis, Anastasios
    MACHINE LEARNING, 2024, 113 (06) : 3569 - 3587
  • [42] Deep Neural Networks With Confidence Sampling For Electrical Anomaly Detection
    Tasfi, Norman L.
    Higashino, Wilson A.
    Grolinger, Katarina
    Capretz, Miriam A. M.
    2017 IEEE INTERNATIONAL CONFERENCE ON INTERNET OF THINGS (ITHINGS) AND IEEE GREEN COMPUTING AND COMMUNICATIONS (GREENCOM) AND IEEE CYBER, PHYSICAL AND SOCIAL COMPUTING (CPSCOM) AND IEEE SMART DATA (SMARTDATA), 2017, : 1038 - 1045
  • [43] Angular sampling reduction in SPECT using deep neural networks
    Kang, Seung Kwan
    Shin, Seong A.
    Lee, Jae Sung
    JOURNAL OF NUCLEAR MEDICINE, 2018, 59
  • [44] Better schedules for low precision training of deep neural networks
    Cameron R. Wolfe
    Anastasios Kyrillidis
    Machine Learning, 2024, 113 : 3569 - 3587
  • [45] Deep learning with coherent VCSEL neural networks
    Zaijun Chen
    Alexander Sludds
    Ronald Davis
    Ian Christen
    Liane Bernstein
    Lamia Ateshian
    Tobias Heuser
    Niels Heermeier
    James A. Lott
    Stephan Reitzenstein
    Ryan Hamerly
    Dirk Englund
    Nature Photonics, 2023, 17 : 723 - 730
  • [46] Learning deep neural networks for node classification
    Li, Bentian
    Pi, Dechang
    EXPERT SYSTEMS WITH APPLICATIONS, 2019, 137 : 324 - 334
  • [47] Inspecting the behaviour of Deep Learning Neural Networks
    Duer, Alexander
    Filzmoser, Peter
    Rauber, Andreas
    ERCIM NEWS, 2019, (116): : 18 - 19
  • [48] Piecewise linear neural networks and deep learning
    Qinghua Tao
    Li Li
    Xiaolin Huang
    Xiangming Xi
    Shuning Wang
    Johan A. K. Suykens
    Nature Reviews Methods Primers, 2
  • [49] Abstraction Hierarchy in Deep Learning Neural Networks
    Ilin, Roman
    Watson, Thomas
    Kozma, Robert
    2017 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2017, : 768 - 774
  • [50] Deep Neural Networks for Learning Graph Representations
    Cao, Shaosheng
    Lu, Wei
    Xu, Qiongkai
    THIRTIETH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2016, : 1145 - 1152