Multilayer Sparsity-Based Tensor Decomposition for Low-Rank Tensor Completion

被引:125
|
作者
Xue, Jize [1 ,2 ]
Zhao, Yongqiang [1 ]
Huang, Shaoguang [3 ]
Liao, Wenzhi [2 ,4 ]
Chan, Jonathan Cheung-Wai [5 ]
Kong, Seong G. [6 ]
机构
[1] Northwestern Polytech Univ Shenzhen, Res & Dev Inst, Shenzhen 518057, Peoples R China
[2] Univ Ghent, IMEC Res Grp, B-9000 Ghent, Belgium
[3] Univ Ghent, Dept Telecommun & Informat Proc TELIN, Grp Artificial Intelligence & Sparse Modeling GAI, B-9000 Ghent, Belgium
[4] Univ Ghent, Flemish Inst Technol Res VITO, B-9000 Ghent, Belgium
[5] Vrije Univ Brussel, Dept Elect & Informat, B-1050 Brussels, Belgium
[6] Sejong Univ, Dept Comp Engn, Seoul 05006, South Korea
基金
中国国家自然科学基金;
关键词
Tensors; Matrix decomposition; Correlation; Nonhomogeneous media; Minimization; Transmission line measurements; Biomedical measurement; CANDECOMP; PARAFAC (CP) decomposition; factor smooth prior; low-rank tensor completion (LRTC); multilayer sparsity (MLS) constraints; subspace structured sparsity; RECOVERY; FACTORIZATION; MINIMIZATION; REGULARIZATION; REPRESENTATION; MODELS;
D O I
10.1109/TNNLS.2021.3083931
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Existing methods for tensor completion (TC) have limited ability for characterizing low-rank (LR) structures. To depict the complex hierarchical knowledge with implicit sparsity attributes hidden in a tensor, we propose a new multilayer sparsity-based tensor decomposition (MLSTD) for the low-rank tensor completion (LRTC). The method encodes the structured sparsity of a tensor by the multiple-layer representation. Specifically, we use the CANDECOMP/PARAFAC (CP) model to decompose a tensor into an ensemble of the sum of rank-1 tensors, and the number of rank-1 components is easily interpreted as the first-layer sparsity measure. Presumably, the factor matrices are smooth since local piecewise property exists in within-mode correlation. In subspace, the local smoothness can be regarded as the second-layer sparsity. To describe the refined structures of factor/subspace sparsity, we introduce a new sparsity insight of subspace smoothness: a self-adaptive low-rank matrix factorization (LRMF) scheme, called the third-layer sparsity. By the progressive description of the sparsity structure, we formulate an MLSTD model and embed it into the LRTC problem. Then, an effective alternating direction method of multipliers (ADMM) algorithm is designed for the MLSTD minimization problem. Various experiments in RGB images, hyperspectral images (HSIs), and videos substantiate that the proposed LRTC methods are superior to state-of-the-art methods.
引用
收藏
页码:6916 / 6930
页数:15
相关论文
共 50 条
  • [1] Robust Low-Rank and Sparse Tensor Decomposition for Low-Rank Tensor Completion
    Shi, Yuqing
    Du, Shiqiang
    Wang, Weilan
    [J]. PROCEEDINGS OF THE 33RD CHINESE CONTROL AND DECISION CONFERENCE (CCDC 2021), 2021, : 7138 - 7143
  • [2] Enhanced Sparsity Prior Model for Low-Rank Tensor Completion
    Xue, Jize
    Zhao, Yongqiang
    Liao, Wenzhi
    Chan, Jonathan Cheung-Wai
    Kong, Seong G.
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2020, 31 (11) : 4567 - 4581
  • [3] Tensor Factorization for Low-Rank Tensor Completion
    Zhou, Pan
    Lu, Canyi
    Lin, Zhouchen
    Zhang, Chao
    [J]. IEEE TRANSACTIONS ON IMAGE PROCESSING, 2018, 27 (03) : 1152 - 1163
  • [4] Tensor Completion using Low-Rank Tensor Train Decomposition by Riemannian Optimization
    Wang, Junli
    Zhao, Guangshe
    Wang, Dingheng
    Li, Guoqi
    [J]. 2019 CHINESE AUTOMATION CONGRESS (CAC2019), 2019, : 3380 - 3384
  • [5] Low-Rank Tensor Completion by Approximating the Tensor Average Rank
    Wang, Zhanliang
    Dong, Junyu
    Liu, Xinguo
    Zeng, Xueying
    [J]. 2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 4592 - 4600
  • [6] Iterative tensor eigen rank minimization for low-rank tensor completion
    Su, Liyu
    Liu, Jing
    Tian, Xiaoqing
    Huang, Kaiyu
    Tan, Shuncheng
    [J]. INFORMATION SCIENCES, 2022, 616 : 303 - 329
  • [7] Adaptive Rank Estimation Based Tensor Factorization Algorithm for Low-Rank Tensor Completion
    Liu, Han
    Liu, Jing
    Su, Liyu
    [J]. PROCEEDINGS OF THE 38TH CHINESE CONTROL CONFERENCE (CCC), 2019, : 3444 - 3449
  • [8] Robust Low-Rank Tensor Completion Based on Tensor Ring Rank via,&epsilon
    Li, Xiao Peng
    So, Hing Cheung
    [J]. IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2021, 69 : 3685 - 3698
  • [9] Higher-dimension Tensor Completion via Low-rank Tensor Ring Decomposition
    Yuan, Longhao
    Cao, Jianting
    Zhao, Xuyang
    Wu, Qiang
    Zhao, Qibin
    [J]. 2018 ASIA-PACIFIC SIGNAL AND INFORMATION PROCESSING ASSOCIATION ANNUAL SUMMIT AND CONFERENCE (APSIPA ASC), 2018, : 1071 - 1076
  • [10] Low-Rank tensor completion based on nonconvex regularization
    Su, Xinhua
    Ge, Huanmin
    Liu, Zeting
    Shen, Yanfei
    [J]. SIGNAL PROCESSING, 2023, 212