Convolutional Neural Network with Structural Input for Visual Object Tracking

被引:6
|
作者
Fiaz, Mustansar [1 ]
Mahmood, Arif [2 ]
Jung, Soon Ki [1 ]
机构
[1] Kyungpook Natl Univ, Sch Comp Sci & Engn, Daegu, South Korea
[2] Informat Technol Univ, Dept Comp Sci, Lahore, Pakistan
关键词
Deep learning; convolutional neural network; visual tracking; machine learning;
D O I
10.1145/3297280.3297416
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Numerous deep learning approaches have been applied to visual object tracking owing to their capabilities to leverage huge training data for performance improvement. Most of these approaches have limitations with regard to learning target specific information rich features and therefore observe reduced accuracy in the presence of different challenges such as occlusion, scale variations, rotation and clutter. We proposed a deep neural network that takes input in the form of two stacked patches and regresses both the similarity and the dis-similarity scores in single evaluation. Image patches are concatenated depth-wise and fed to a six channel input of the network. The proposed network is generic and exploits the structural differences between the two input patches to obtain more accurate similarity and dissimilarity scores. Online learning is enforced via short-term and long-term updates to improve the tracking performance. Extensive experimental evaluations have been performed on OTB2015 and TempleColor128 benchmark datasets. Comparisons with state-of-the-art methods indicate that the proposed framework has achieved better tracking performance. The proposed tracking framework has obtained improved accuracy in different challenges including occlusion, background clutter, in-plane rotation and scale variations.
引用
收藏
页码:1345 / 1352
页数:8
相关论文
共 50 条
  • [1] Visual object tracking with adaptive structural convolutional network
    Yuan, Di
    Li, Xin
    He, Zhenyu
    Liu, Qiao
    Lu, Shuwei
    [J]. KNOWLEDGE-BASED SYSTEMS, 2020, 194
  • [2] Visual Object Tracking Based on Bilinear Convolutional Neural Network
    Zhang Chunting
    [J]. LASER & OPTOELECTRONICS PROGRESS, 2020, 57 (04)
  • [3] CONTINUITY-DISCRIMINATION CONVOLUTIONAL NEURAL NETWORK FOR VISUAL OBJECT TRACKING
    Li, Shen
    Ma, Bingpeng
    Chang, Hong
    Shan, Shiguang
    Chen, Xilin
    [J]. 2018 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO (ICME), 2018,
  • [4] Visual Tracking with Convolutional Neural Network
    Zhang, Le
    Suganthan, P. N.
    [J]. 2015 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS (SMC 2015): BIG DATA ANALYTICS FOR HUMAN-CENTRIC SYSTEMS, 2015, : 2072 - 2077
  • [5] Visual tracking based on convolutional neural network
    Li, Xiuzhi
    Jiang, Kai
    Jia, Songmin
    Zhang, Xiangyin
    Sun, Yanjun
    [J]. 2018 CHINESE AUTOMATION CONGRESS (CAC), 2018, : 4061 - 4066
  • [6] Robust Visual Tracking with Deep Convolutional Neural Network based Object Proposals on PETS
    Zhu, Gao
    Porikli, Fatih
    Li, Hongdong
    [J]. PROCEEDINGS OF 29TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS, (CVPRW 2016), 2016, : 1265 - 1272
  • [7] GATE CONNECTED CONVOLUTIONAL NEURAL NETWORK FOR OBJECT TRACKING
    Kokul, T.
    Fookes, C.
    Sridharan, S.
    Ramanan, A.
    Pinidiyaarachchi, U. A. J.
    [J]. 2017 24TH IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2017, : 2602 - 2606
  • [8] A Visual Tracking Deep Convolutional Neural Network Accelerator
    Qin, Zhiyong
    Yu, Lixin
    [J]. PROCEEDINGS OF THE 2017 2ND INTERNATIONAL CONFERENCE ON AUTOMATION, MECHANICAL CONTROL AND COMPUTATIONAL ENGINEERING (AMCCE 2017), 2017, 118 : 493 - 499
  • [9] Detection based visual tracking with convolutional neural network
    Wang, Yong
    Luo, Xinbin
    Ding, Lu
    Fu, Shan
    Wei, Xian
    [J]. KNOWLEDGE-BASED SYSTEMS, 2019, 175 : 62 - 71
  • [10] Object Tracking in the Video Stream by Means of a Convolutional Neural Network
    Zolotukhin, Yu N.
    Kotov, K. Yu
    Nesterov, A. A.
    Semenyuk, E. D.
    [J]. OPTOELECTRONICS INSTRUMENTATION AND DATA PROCESSING, 2020, 56 (06) : 642 - 648