AMFFNet: Asymmetric Multiscale Feature Fusion Network of RGB-NIR for Solid Waste Detection

被引:1
|
作者
Cai, Zhenxing [1 ]
Fang, Huaiying [1 ]
Jiang, Fengfeng [2 ]
Yang, Jianhong [1 ]
Ji, Tianchen [1 ]
Hu, Yangyang [1 ]
Wang, Xin [1 ]
机构
[1] Huaqiao Univ, Fujian Prov Univ, Huaqiao Univ, Sch Electromech & Automat,Key Lab Proc Monitoring, Xiamen 361021, Peoples R China
[2] Xiamen Luhai Environm Protect Co Ltd, Xiamen 361001, Peoples R China
关键词
Asymmetric multiscale feature fusion; hyper-spectral images; multisensor fusion technology; object detection; residual network (ResNet);
D O I
10.1109/TIM.2023.3300445
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Multimodal fusion object detection aims to improve detection accuracy by integrating information from multiple modalities. RGB image-based object detection makes finely sorting solid waste according to material challenging. This article built a dual-camera acquisition platform using a line-scan color camera and a hyperspectral camera to collect RGB and hyperspectral images. In order to use RGB images and hyperspectral images for feature fusion more effectively, we propose an asymmetric multiscale feature fusion network (AMFFNet) based on RGB-near-infrared (NIR) multisensor fusion technology. Specifically, we designed a hyperspectral image convolution unit (HICU) to fully extract multiscale features from hyperspectral images. Second, we concatenate the hyperspectral feature maps and the feature maps output by feature pyramid networks (FPNs) of the RGB image feature extraction stage to achieve asymmetric multiscale feature fusion. In addition, a dimensionality reduction strategy (DRS) is proposed to remove the bands that are redundant and have a low signal-to-noise ratio in hyperspectral images. Ablation studies have confirmed the effectiveness of AMFFNet components, and we have conducted extensive integration experiments on the article-based solid waste dataset. The experimental results and analysis show that AMFFNet based on mask region-based convolutional neural network (Mask RCNN), faster RCNN, and RetinaNet detectors outperforms the original model by 5.05%, 3.57%, and 6.36% on AP at IoU = 0.5 and by 3.1%, 2.05%, and 3.73% on COCO's standard AP metric, and the parameters of the network increased by 0.47-0.56 M. In conclusion, the proposed method effectively improves the performance of the object detection model for fine identification of solid waste.
引用
收藏
页数:10
相关论文
共 50 条
  • [21] MFFENet: Multiscale Feature Fusion and Enhancement Network For RGB-Thermal Urban Road Scene Parsing
    Zhou, Wujie
    Lin, Xinyang
    Lei, Jingsheng
    Yu, Lu
    Hwang, Jenq-Neng
    IEEE Transactions on Multimedia, 2022, 24 : 2526 - 2538
  • [22] A Multiscale and Multidirection Feature Fusion Network for Road Detection From Satellite Imagery
    Wang, Yuchuan
    Tong, Ling
    Luo, Shiyu
    Xiao, Fanghong
    Yang, Jiaxing
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2024, 62 : 1 - 18
  • [23] Global attention network with multiscale feature fusion for infrared small target detection
    Zhang, Fan
    Lin, Shunlong
    Xiao, Xiaoyang
    Wang, Yun
    Zhao, Yuqian
    OPTICS AND LASER TECHNOLOGY, 2024, 168
  • [24] A Non-Local Attention Feature Fusion Network for Multiscale Object Detection
    Wu, Xuke
    Xiong, Gang
    Tian, Bin
    Song, Bing
    Lu, Bo
    Liu, Sheng
    Zhu, Fenghua
    IEEE JOURNAL OF RADIO FREQUENCY IDENTIFICATION, 2022, 6 : 733 - 738
  • [25] Object Detection For Remote Sensing Image Based on Multiscale Feature Fusion Network
    Tian Tingting
    Yang Jun
    LASER & OPTOELECTRONICS PROGRESS, 2022, 59 (16)
  • [26] Multiscale network based on feature fusion for fire disaster detection in complex scenes
    Feng, Jian
    Sun, Yu
    EXPERT SYSTEMS WITH APPLICATIONS, 2024, 240
  • [27] MFFNet: Multimodal feature fusion network for RGB-D transparent object detection
    Zhu, Li
    Li, Tuanjie
    Ning, Yuming
    Zhang, Yan
    INTERNATIONAL JOURNAL OF ADVANCED ROBOTIC SYSTEMS, 2024, 21 (05):
  • [28] AMCFNet: Asymmetric multiscale and crossmodal fusion network for RGB-D semantic segmentation in indoor service robots
    Zhou, Wujie
    Yue, Yuchun
    Fang, Meixin
    Mao, Shanshan
    Yang, Rongwang
    Yu, Lu
    JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION, 2023, 97
  • [29] Asymmetric Adaptive Fusion in a Two-Stream Network for RGB-D Human Detection
    Zhang, Wenli
    Guo, Xiang
    Wang, Jiaqi
    Wang, Ning
    Chen, Kaizhen
    SENSORS, 2021, 21 (03) : 1 - 17
  • [30] SAR Ship Detection Based on Convolutional Neural Network with Deep Multiscale Feature Fusion
    Long, Yang
    Juan, Su
    Hua, Huang
    Xiang, Li
    ACTA OPTICA SINICA, 2020, 40 (02)