INFNet :Deep instance feature chain learning network for panoptic segmentation

被引:0
|
作者
Mao L. [1 ]
Ren F.-Z. [1 ]
Yang D.-W. [1 ]
Zhang R.-B. [1 ]
机构
[1] School of Electromechanical Engineering, Dalian Minzu University, Dalian
关键词
Chain network; Edge feature; Instance feature; Panoptic segmentation; Shortcut connection;
D O I
10.37188/OPE.20202812.2665
中图分类号
学科分类号
摘要
A novel deep instance feature chain learning network for panoptic segmentation (INFNet) was developed to solve the problem of failure of target boundary segmentation caused by insufficient instant feature extraction in panoptic segmentation. This network consisted of a basic chain unit, whose functions were divided into two types, feature holding chain and feature enhancement chain, based on the different methods of processing feature information by the unit structure. The feature-holding chain represented the input stage of the extraction of a chain network feature, in which the integrity of the input information was guaranteed, and then this feature was transmitted to the feature-enhancement chain structure. The feature-enhancement chain increased the network depth and improved the feature extraction ability through its extension. INFNet could obtain adequate edge feature information and improve segmentation accuracy, owing to the robust depth-stacking characteristics. The experiment results for the MS COCO and Cityscapes datasets showed that our INFNet was superior to similar existing methods in terms of segmentation accuracy. Compared to the Mask RCNN instance segmentation structure widely used in panoptic segmentation networks, the segmentation accuracy of INFNet increased by up to 0.94%. © 2020, Science Press. All right reserved.
引用
下载
收藏
页码:2665 / 2673
页数:8
相关论文
共 21 条
  • [1] HE K, GKIOXARI G, DOLLAR P, Et al., Mask R-CNN, IEEE International Conference on Computer Vision, pp. 2980-2988, (2017)
  • [2] HE K, ZHANG X, REN S, Et al., Deep Residual Learning for Image Recognition, IEEE Conference on Computer Vision and Pattern Recognition, pp. 770-778, (2016)
  • [3] KIRILLOV A, GIRSHICK R, HE K, Et al., Panoptic Feature Pyramid Networks, IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 6392-6401, (2019)
  • [4] LONG J, SHELHAMER E, DARRELL T., Fully convolutional networks for semantic segmentation, IEEE Conference on Computer Vision and Pattern Recognition, pp. 3431-3440, (2015)
  • [5] CHEN L C, GEORGE PAPANDREOU, IASONAS KOKKINOS, Et al., DeepLab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected CRFs, IEEE Transactions on Pattern Analysis and Machine Intelligence, 40, 4, pp. 834-848, (2018)
  • [6] REN F L, HE X, WEI ZH H, Et al., Semantic segmentation based on DeepLabV3+ and superpixel optimization, Opt. Precision Eng, 27, 12, pp. 2722-2729, (2019)
  • [7] BADRINARAYANAN V, KENDALL A, CIPOLLA R., SegNet: A deep convolutional encoder-decoder architecture for image segmentation, IEEE Transactions on Pattern Analysis and Machine Intelligence, 39, 12, pp. 2481-2495, (2017)
  • [8] XIONG Y, LIAO R, ZHAO H, Et al., UPSNet: A Unified Panoptic Segmentation Network, IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 8810-8818, (2019)
  • [9] LIU H, PENG C, YU C, Et al., An End-To-End Network for Panoptic Segmentation, IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 6165-6174, (2019)
  • [10] LI J, RAVENTOS A, BHARGAVA A, Et al., Learning to fuse things and stuff