Learning a Better Negative Sampling Policy with Deep Neural Networks for Search

被引:6
|
作者
Cohen, Daniel [1 ]
Jordan, Scott M. [2 ]
Croft, W. Bruce [1 ]
机构
[1] Univ Massachusetts Amherst, Ctr Intelligent Informat Retrieval, Amherst, MA 01003 USA
[2] Univ Massachusetts Amherst, Autonomous Learning Lab, Amherst, MA 01003 USA
关键词
D O I
10.1145/3341981.3344220
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
In information retrieval, sampling methods used to select documents for neural models must often deal with large class imbalances during training. This issue necessitates careful selection of negative instances when training neural models to avoid the risk of overfitting. For most work, heuristic sampling approaches, or policies, are created based off of domain experts, such as choosing samples with high BM25 scores or a random process over candidate documents. However, these sampling approaches are done with the test distribution in mind. In this paper, we demonstrate that the method chosen to sample negative documents during training plays a critical role in both the stability of training, as well as overall performance. Furthermore, we establish that using reinforcement learning to optimize a policy over a set of sampling functions can significantly improve performance over standard training practices with respect to IR metrics and is robust to hyperparameters and random seeds.
引用
收藏
页码:19 / 26
页数:8
相关论文
共 50 条
  • [31] Multiplierless Neural Networks for Deep Learning
    Banduka, Maja Lutovac
    Lutovac, Miroslav
    2024 13TH MEDITERRANEAN CONFERENCE ON EMBEDDED COMPUTING, MECO 2024, 2024, : 262 - 265
  • [32] Shortcut learning in deep neural networks
    Geirhos, Robert
    Jacobsen, Joern-Henrik
    Michaelis, Claudio
    Zemel, Richard
    Brendel, Wieland
    Bethge, Matthias
    Wichmann, Felix A.
    NATURE MACHINE INTELLIGENCE, 2020, 2 (11) : 665 - 673
  • [33] Towards Better Analysis of Deep Convolutional Neural Networks
    Liu, Mengchen
    Shi, Jiaxin
    Li, Zhen
    Li, Chongxuan
    Zhu, Jun
    Liu, Shixia
    IEEE TRANSACTIONS ON VISUALIZATION AND COMPUTER GRAPHICS, 2017, 23 (01) : 91 - 100
  • [34] Neural Architecture Search Using Deep Neural Networks and Monte Carlo Tree Search
    Wang, Linnan
    Zhao, Yiyang
    Yuu Jinnai
    Tian, Yuandong
    Fonseca, Rodrigo
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 9983 - 9991
  • [35] Deep Learning Neural Networks and Bayesian Neural Networks in Data Analysis
    Chernoded, Andrey
    Dudko, Lev
    Myagkov, Igor
    Volkov, Petr
    XXIII INTERNATIONAL WORKSHOP HIGH ENERGY PHYSICS AND QUANTUM FIELD THEORY (QFTHEP 2017), 2017, 158
  • [36] Introduction to Machine Learning, Neural Networks, and Deep Learning
    Choi, Rene Y.
    Coyner, Aaron S.
    Kalpathy-Cramer, Jayashree
    Chiang, Michael F.
    Campbell, J. Peter
    TRANSLATIONAL VISION SCIENCE & TECHNOLOGY, 2020, 9 (02):
  • [37] DeepHyper: Asynchronous Hyperparameter Search for Deep Neural Networks
    Balaprakash, Prasanna
    Salim, Michael
    Uram, Thomas D.
    Vishwanath, Venkat
    Wild, Stefan M.
    2018 IEEE 25TH INTERNATIONAL CONFERENCE ON HIGH PERFORMANCE COMPUTING (HIPC), 2018, : 42 - 51
  • [38] Deep Metric Learning Using Negative Sampling Probability Annealing
    Kertesz, Gabor
    SENSORS, 2022, 22 (19)
  • [39] Exploring Neural Architecture Search Space via Deep Deterministic Sampling
    Mills, Keith G.
    Salameh, Mohammad
    Niu, Di
    Han, Fred X.
    Rezaei, Seyed Saeed Changiz
    Yao, Hengshuai
    Lu, Wei
    Lian, Shuo
    Jui, Shangling
    IEEE ACCESS, 2021, 9 : 110962 - 110974
  • [40] Automated Deep Learning: Neural Architecture Search Is Not the End
    Dong, Xuanyi
    Kedziora, David Jacob
    Musial, Katarzyna
    Gabrys, Bogdan
    FOUNDATIONS AND TRENDS IN MACHINE LEARNING, 2024, 17 (05): : 767 - 920