Consistent Hierarchical Classification with A Generalized Metric

被引:0
|
作者
Cao, Yuzhou [1 ]
Feng, Lei [1 ]
An, Bo [1 ,2 ]
机构
[1] Nanyang Technol Univ, Sch Comp Sci & Engn, Singapore, Singapore
[2] Skywork AI, Singapore, Singapore
来源
INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 238 | 2024年 / 238卷
基金
新加坡国家研究基金会;
关键词
MULTICLASS CLASSIFICATION;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In multi-class hierarchical classification, a natural evaluation metric is the tree distance loss that takes the value of two labels' distance on the pre-defined tree hierarchy. This metric is motivated by that its Bayes optimal solution is the deepest label on the tree whose induced superclass (subtree rooted at it) includes the true label with probability at least 12. However, it can hardly handle the risk sensitivity of different tasks since its accuracy requirement for induced superclasses is fixed at 12. In this paper, we first introduce a new evaluation metric that generalizes the tree distance loss, whose solution's accuracy constraint 1+c/2 can be controlled by a penalty value c tailored for different tasks: a higher c indicates the emphasis on prediction's accuracy and a lower one indicates that on specificity. Then, we propose a novel class of consistent surrogate losses based on an intuitive presentation of our generalized metric and its regret, which can be compatible with various binary losses. Finally, we theoretically derive the regret transfer bounds for our proposed surrogates and empirically validate their usefulness on benchmark datasets.
引用
收藏
页数:16
相关论文
共 50 条