Optimizing non-decomposable measures with deep networks

被引:0
|
作者
Amartya Sanyal
Pawan Kumar
Purushottam Kar
Sanjay Chawla
Fabrizio Sebastiani
机构
[1] The University of Oxford,
[2] The Alan Turing Institute,undefined
[3] Indian Institute of Technology Kanpur,undefined
[4] Qatar Computing Research Institute,undefined
[5] Istituto di Scienza e Tecnologia dell’Informazione,undefined
来源
Machine Learning | 2018年 / 107卷
关键词
Optimization; Deep learning; F-measure; Task-specific training;
D O I
暂无
中图分类号
学科分类号
摘要
We present a class of algorithms capable of directly training deep neural networks with respect to popular families of task-specific performance measures for binary classification such as the F-measure, QMean and the Kullback–Leibler divergence that are structured and non-decomposable. Our goal is to address tasks such as label-imbalanced learning and quantification. Our techniques present a departure from standard deep learning techniques that typically use squared or cross-entropy loss functions (that are decomposable) to train neural networks. We demonstrate that directly training with task-specific loss functions yields faster and more stable convergence across problems and datasets. Our proposed algorithms and implementations offer several advantages including (i) the use of fewer training samples to achieve a desired level of convergence, (ii) a substantial reduction in training time, (iii) a seamless integration of our implementation into existing symbolic gradient frameworks, and (iv) assurance of convergence to first order stationary points. It is noteworthy that the algorithms achieve this, especially point (iv), despite being asked to optimize complex objective functions. We implement our techniques on a variety of deep architectures including multi-layer perceptrons and recurrent neural networks and show that on a variety of benchmark and real data sets, our algorithms outperform traditional approaches to training deep networks, as well as popular techniques used to handle label imbalance.
引用
收藏
页码:1597 / 1620
页数:23
相关论文
共 50 条
  • [1] Optimizing non-decomposable measures with deep networks
    Sanyal, Amartya
    Kumar, Pawan
    Kar, Purushottam
    Chawla, Sanjay
    Sebastiani, Fabrizio
    MACHINE LEARNING, 2018, 107 (8-10) : 1597 - 1620
  • [2] Optimizing Non-decomposable Performance Measures: A Tale of Two Classes
    Narasimhan, Harikrishna
    Kar, Purushottam
    Jain, Prateek
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 37, 2015, 37 : 199 - 208
  • [3] Simple Weak Coresets for Non-decomposable Classification Measures
    Malaviya, Jayesh
    Dasgupta, Anirban
    Chhaya, Rachit
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 13, 2024, : 14289 - 14296
  • [4] Optimizing Non-Decomposable Evaluation Metrics for Neural Machine Translation
    Shi-Qi Shen
    Yang Liu
    Mao-Song Sun
    Journal of Computer Science and Technology, 2017, 32 : 796 - 804
  • [5] Optimizing Non-Decomposable Evaluation Metrics for Neural Machine Translation
    Shen, Shi-Qi
    Liu, Yang
    Sun, Mao-Song
    JOURNAL OF COMPUTER SCIENCE AND TECHNOLOGY, 2017, 32 (04) : 796 - 804
  • [6] NON-DECOMPOSABLE NAMBU BRACKETS
    Bering, Klaus
    ARCHIVUM MATHEMATICUM, 2015, 51 (04): : 211 - 232
  • [7] Fair learning withWasserstein barycenters for non-decomposable performance measures
    Gaucher, Solenne
    Schreuder, Nicolas
    Chzhen, Evgenii
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 206, 2023, 206
  • [8] Multiclass Learning from Noisy Labels for Non-decomposable Performance Measures
    Zhang, Mingyuan
    Agarwal, Shivani
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 238, 2024, 238
  • [9] On the Statistical Consistency of Plug-in Classifiers for Non-decomposable Performance Measures
    Narasimhan, Harikrishna
    Vaish, Rohit
    Agarwal, Shivani
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 27 (NIPS 2014), 2014, 27
  • [10] Cost-Sensitive Self-Training for Optimizing Non-Decomposable Metrics
    Rangwani, Harsh
    Ramasubramanian, Shrinivas
    Takemori, Sho
    Kato, Takashi
    Umeda, Yuhei
    Babu, R. Venkatesh
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,