A tensor compression algorithm using Tucker decomposition and dictionary dimensionality reduction

被引:2
|
作者
Gan, Chenquan [1 ,2 ,3 ]
Mao, Junwei [1 ,2 ,3 ]
Zhang, Zufan [1 ,2 ,3 ]
Zhu, Qingyi [4 ]
机构
[1] Chongqing Univ Posts & Telecommun, Sch Commun & Informat Engn, Chongqing 400065, Peoples R China
[2] Chongqing Key Lab Mobile Commun Technol, Chongqing, Peoples R China
[3] Minist Educ, Engn Res Ctr Mobile Commun, Chongqing, Peoples R China
[4] Chongqing Univ Posts & Telecommun, Sch Cyber Secur & Informat Law, Chongqing, Peoples R China
关键词
Tensor signal compression; Tucker decomposition; sparse representation; dictionary learning; denoising ability;
D O I
10.1177/1550147720916408
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Tensor compression algorithms play an important role in the processing of multidimensional signals. In previous work, tensor data structures are usually destroyed by vectorization operations, resulting in information loss and new noise. To this end, this article proposes a tensor compression algorithm using Tucker decomposition and dictionary dimensionality reduction, which mainly includes three parts: tensor dictionary representation, dictionary preprocessing, and dictionary update. Specifically, the tensor is respectively performed by the sparse representation and Tucker decomposition, from which one can obtain the dictionary, sparse coefficient, and core tensor. Furthermore, the sparse representation can be obtained through the relationship between sparse coefficient and core tensor. In addition, the dimensionality of the input tensor is reduced by using the concentrated dictionary learning. Finally, some experiments show that, compared with other algorithms, the proposed algorithm has obvious advantages in preserving the original data information and denoising ability.
引用
收藏
页数:10
相关论文
共 50 条
  • [1] Dimensionality reduction algorithm of tensor data based on orthogonal tucker decomposition and local discrimination difference
    Gao, Wenxu
    Ma, Zhengming
    Yuan, Xuejing
    APPLIED INTELLIGENCE, 2022, 52 (12) : 14518 - 14540
  • [2] Dimensionality reduction algorithm of tensor data based on orthogonal tucker decomposition and local discrimination difference
    Wenxu Gao
    Zhengming Ma
    Xuejing Yuan
    Applied Intelligence, 2022, 52 : 14518 - 14540
  • [3] TENSOR DICTIONARY LEARNING WITH SPARSE TUCKER DECOMPOSITION
    Zubair, Syed
    Wang, Wenwu
    2013 18TH INTERNATIONAL CONFERENCE ON DIGITAL SIGNAL PROCESSING (DSP), 2013,
  • [4] DICTIONARY LEARNING HYPERSPECTRAL TARGET DETECTION ALGORITHM BASED ON TUCKER TENSOR DECOMPOSITION
    Zhao, Chunhui
    Wang, Mingxing
    Su, Nan
    Feng, Shou
    IGARSS 2020 - 2020 IEEE INTERNATIONAL GEOSCIENCE AND REMOTE SENSING SYMPOSIUM, 2020, : 1763 - 1766
  • [5] Research of Incremental Dimensionality Reduction Based on Tensor Decomposition Algorithm
    Guo, Xin
    Xiang, Yang
    Lv, Dongdong
    Yuan, Shuhan
    Huang, Yinfei
    Zhang, Qi
    Wang, Jisheng
    Wang, Dong
    WIRELESS COMMUNICATIONS, NETWORKING AND APPLICATIONS, WCNA 2014, 2016, 348 : 87 - 94
  • [6] Hyperspectral image, video compression using sparse tucker tensor decomposition
    Das, Samiran
    IET IMAGE PROCESSING, 2021, 15 (04) : 964 - 973
  • [7] Cross Tensor Approximation Methods for Compression and Dimensionality Reduction
    Ahmadi-Asl, Salman
    Caiafa, Cesar F.
    Cichocki, Andrzej
    Phan, Anh Huy
    Tanaka, Toshihisa
    Oseledets, Ivan
    Wang, Jun
    IEEE ACCESS, 2021, 9 : 150809 - 150838
  • [8] Spectrum Map Construction Algorithm Based on Tensor Tucker Decomposition
    Chen Z.
    Hu J.
    Zhang B.
    Guo D.
    Dianzi Yu Xinxi Xuebao/Journal of Electronics and Information Technology, 2023, 45 (11): : 4161 - 4169
  • [9] Tucker Tensor Decomposition on FPGA
    Zhang, Kaiqi
    Zhang, Xiyuan
    Zhang, Zheng
    2019 IEEE/ACM INTERNATIONAL CONFERENCE ON COMPUTER-AIDED DESIGN (ICCAD), 2019,
  • [10] Improved non-negative tensor Tucker decomposition algorithm for interference hyper-spectral image compression
    WEN Jia
    ZHAO JunSuo
    MA CaiWen
    WANG CaiLing
    Science China(Information Sciences), 2015, 58 (05) : 112 - 120