CephaNN: A Multi-Head Attention Network for Cephalometric Landmark Detection

被引:27
|
作者
Qian, Jiahong [1 ]
Luo, Weizhi [1 ]
Cheng, Ming [1 ]
Tao, Yubo [1 ,2 ]
Lin, Jun [3 ]
Lin, Hai [1 ,2 ]
机构
[1] Zhejiang Univ, State Key Lab CAD&CG, Hangzhou 310058, Peoples R China
[2] Zhejiang Univ, Innovat Ctr Minimally Invas Tech & Device, Hangzhou 310058, Peoples R China
[3] Zhejiang Univ, Coll Med, Affiliated Hosp 1, Dept Stomatol, Hangzhou 310058, Peoples R China
基金
中国国家自然科学基金;
关键词
Heating systems; Neural networks; Kernel; Feature extraction; Annotations; Two dimensional displays; Deep learning; Cephalometric landmark detection; multi-head attention; neural network; intermediate supervision; region enhance;
D O I
10.1109/ACCESS.2020.3002939
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Cephalometric landmark detection is a crucial step in orthodontic and orthognathic treatments. To detect cephalometric landmarks accurately, we propose a novel multi-head attention neural network (CephaNN). CephaNN is an end-to-end network based on the heatmaps of annotated landmarks, and it consists of two parts, the multi-head part and the attention part. In the multi-head part, we adopt multi-head subnets to gain comprehensive knowledge of various subspaces of a cephalogram. The intermediate supervision is applied to accelerate the convergence. Based on the feature maps learned from the multi-head Part, the attention part applies the multi-attention mechanism to obtain a refined detection. For solving the class imbalance problem, we propose a region enhancing (RE) loss, to enhance the efficient regions on the regressed heatmaps. Experiments in the benchmark dataset demonstrate that CephaNN is state-of-the-art with the detection accuracy of 87.61% in the clinically accepted 2.0-mm range. Furthermore, CephaNN is efficient in classifying the anatomical types and robust in a real application on a 75-landmark dataset.
引用
收藏
页码:112633 / 112641
页数:9
相关论文
共 50 条
  • [21] Bimodal Fusion Network with Multi-Head Attention for Multimodal Sentiment Analysis
    Zhang, Rui
    Xue, Chengrong
    Qi, Qingfu
    Lin, Liyuan
    Zhang, Jing
    Zhang, Lun
    APPLIED SCIENCES-BASEL, 2023, 13 (03):
  • [22] A Dual Multi-Head Contextual Attention Network for Hyperspectral Image Classification
    Liang, Miaomiao
    He, Qinghua
    Yu, Xiangchun
    Wang, Huai
    Meng, Zhe
    Jiao, Licheng
    REMOTE SENSING, 2022, 14 (13)
  • [23] Acoustic Word Embedding Based on Multi-Head Attention Quadruplet Network
    Zhu, Shirong
    Zhang, Ying
    He, Kai
    Zhao, Lasheng
    IEEE SIGNAL PROCESSING LETTERS, 2022, 29 : 184 - 188
  • [24] Siamese Network cooperating with Multi-head Attention for semantic sentence matching
    Yuan, Zhao
    Jun, Sun
    2020 19TH INTERNATIONAL SYMPOSIUM ON DISTRIBUTED COMPUTING AND APPLICATIONS FOR BUSINESS ENGINEERING AND SCIENCE (DCABES 2020), 2020, : 235 - 238
  • [25] Personalized multi-head self-attention network for news recommendation
    Zheng, Cong
    Song, Yixuan
    NEURAL NETWORKS, 2025, 181
  • [26] Distract Your Attention: Multi-Head Cross Attention Network for Facial Expression Recognition
    Wen, Zhengyao
    Lin, Wenzhong
    Wang, Tao
    Xu, Ge
    BIOMIMETICS, 2023, 8 (02)
  • [27] Capsule Network Improved Multi-Head Attention for Word Sense Disambiguation
    Cheng, Jinfeng
    Tong, Weiqin
    Yan, Weian
    APPLIED SCIENCES-BASEL, 2021, 11 (06):
  • [28] Multi-Head Spatiotemporal Attention Graph Convolutional Network for Traffic Prediction
    Oluwasanmi, Ariyo
    Aftab, Muhammad Umar
    Qin, Zhiguang
    Sarfraz, Muhammad Shahzad
    Yu, Yang
    Rauf, Hafiz Tayyab
    SENSORS, 2023, 23 (08)
  • [29] Multi-Head Attention Neural Network for Smartphone Invariant Indoor Localization
    Tiku, Saideep
    Gufran, Danish
    Pasricha, Sudeep
    2022 IEEE 12TH INTERNATIONAL CONFERENCE ON INDOOR POSITIONING AND INDOOR NAVIGATION (IPIN 2022), 2022,
  • [30] Gaze Estimation Network Based on Multi-Head Attention, Fusion, and Interaction
    Li, Changli
    Li, Fangfang
    Zhang, Kao
    Chen, Nenglun
    Pan, Zhigeng
    SENSORS, 2025, 25 (06)