Ontology-aware Learning and Evaluation for Audio Tagging

被引:0
|
作者
Liu, Haohe [1 ]
Kong, Qiuqiang [2 ]
Liu, Xubo [1 ]
Mei, Xinhao [1 ]
Wang, Wenwu [1 ]
Plumbley, Mark D. [1 ]
机构
[1] Univ Surrey, CVSSP, Guildford, Surrey, England
[2] ByteDance, Speech Audio & Mus Intelligence SAMI Grp, Beijing, Peoples R China
来源
关键词
machine learning; audio tagging; ontology; evaluation metric; CLASSIFICATION;
D O I
10.21437/Interspeech.2023-979
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
This study defines a new evaluation metric for audio tagging tasks to alleviate the limitation of the mean average precision (mAP) metric. The mAP metric treats different kinds of sound as independent classes without considering their relations. The proposed metric, ontology-aware mean average precision (OmAP), addresses the weaknesses of mAP by utilizing additional ontology during evaluation. Specifically, we reweight the false positive events in the model prediction based on the AudioSet ontology graph distance to the target classes. The OmAP also provides insights into model performance by evaluating different coarse-grained levels in the ontology graph. We conduct a human assessment and show that OmAP is more consistent with human perception than mAP. We also propose an ontology-based loss function (OBCE) that reweights binary cross entropy (BCE) loss based on the ontology distance. Our experiment shows that OBCE can improve both mAP and OmAP metrics on the AudioSet tagging task.
引用
收藏
页码:3799 / 3803
页数:5
相关论文
共 50 条
  • [31] OPAL: Ontology-Aware Pretrained Language Model for End-to-End Task-Oriented Dialogue
    Chen, Zhi
    Liu, Yuncong
    Chen, Lu
    Zhu, Su
    Wu, Mengyue
    Yu, Kai
    TRANSACTIONS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, 2023, 11 : 68 - 84
  • [32] Ontology-aware deep learning enables ultrafast and interpretable source tracking among sub-million microbial community samples from hundreds of niches
    Yuguo Zha
    Hui Chong
    Hao Qiu
    Kai Kang
    Yuzheng Dun
    Zhixue Chen
    Xuefeng Cui
    Kang Ning
    Genome Medicine, 14
  • [33] Ontology-aware deep learning enables ultrafast and interpretable source tracking among sub-million microbial community samples from hundreds of niches
    Zha, Yuguo
    Chong, Hui
    Qiu, Hao
    Kang, Kai
    Dun, Yuzheng
    Chen, Zhixue
    Cui, Xuefeng
    Ning, Kang
    GENOME MEDICINE, 2022, 14 (01)
  • [34] An Ontology-Aware Model-Driven Approach for Service-Oriented Application Development: A Stepwise Refinement Manner
    Pourali, Abdolghader
    Abadeh, Maryam Nooraei
    INTERNATIONAL JOURNAL OF INFORMATION TECHNOLOGY & DECISION MAKING, 2024,
  • [35] Ontology-aware classification of tissue and cell-type signals in gene expression profiles across platforms and technologies
    Lee, Young-suk
    Krishnan, Arjun
    Zhu, Qian
    Troyanskaya, Olga G.
    BIOINFORMATICS, 2013, 29 (23) : 3036 - 3044
  • [36] An ontology-aware integration of clinical models, terminologies and guidelines: an exploratory study of the Scale for the Assessment and Rating of Ataxia (SARA)
    Haitham Maarouf
    María Taboada
    Hadriana Rodriguez
    Manuel Arias
    Ángel Sesar
    María Jesús Sobrido
    BMC Medical Informatics and Decision Making, 17
  • [37] An ontology-aware integration of clinical models, terminologies and guidelines: an exploratory study of the Scale for the Assessment and Rating of Ataxia (SARA)
    Maarouf, Haitham
    Taboada, Maria
    Rodriguez, Hadriana
    Arias, Manuel
    Sesar, Angel
    Jesus Sobrido, Maria
    BMC MEDICAL INFORMATICS AND DECISION MAKING, 2017, 17
  • [38] Reinforcement Learning based Neural Architecture Search for Audio Tagging
    Liu, Haiyang
    Zhang, Cheng
    2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
  • [39] Tag Taxonomy Aware Dictionary Learning for Region Tagging
    Zheng, Jingjing
    Jiang, Zhuolin
    2013 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2013, : 369 - 376
  • [40] Enhanced Feature Learning with Normalized Knowledge Distillation for Audio Tagging
    Tang, Yuwu
    Ma, Ziang
    Zhang, Haitao
    INTERSPEECH 2024, 2024, : 1695 - 1699