Multimodal Decoupled Representation With Compatibility Learning for Explicit Nonstationary Process Monitoring

被引:3
|
作者
Song, Pengyu [1 ]
Zhao, Chunhui [1 ]
Ding, Jinliang [2 ]
Zhao, Shunyi [3 ]
机构
[1] Zhejiang Univ, Coll Control Sci & Engn, Hangzhou 310027, Peoples R China
[2] Northeastern Univ, State Key Lab Synthet Automat Proc Ind, Shenyang 110819, Peoples R China
[3] Jiangnan Univ, Key Lab Adv Proc Controlfor Light Ind, Minist Educ, Wuxi 214122, Peoples R China
基金
国家重点研发计划;
关键词
Cascaded learning architecture; compat-ibility learning; explicit process monitoring; information refinement; multimodal decoupled representation; nonstationarity; NEURAL-NETWORKS; COINTEGRATION; STATIONARY;
D O I
10.1109/TIE.2023.3299013
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Frequent switching of operating conditions in industrial processes tends to make the data distribution time varying and variable correlations nonuniform, which brings considerable challenges for explicit representation and monitoring of nonstationary processes. This study addresses the abovementioned problem based on the following recognitions: 1) despite changes in operating conditions, there exist time-invariant process mechanisms, which are nonlinearly coupled with nonstationarity; 2) such a coupling relationship is not uniform but may show diverse modes changing with time, defined as multimodal coupling here; 3) diverse coupling relations can be derived from the superposition of nonstationarity induced by definite driving forces, i.e., changeable operating condition settings, reflecting their intrinsic association under complex changes. Thereupon, a cascaded deep information separation (CDIS) architecture with a compatibility learning algorithm is proposed to extract multimodal decoupled representations. We design an information refinement module (IRM) to capture the basic coupling source (BCS) under the influence of driving forces, where a shortcut connection is incorporated into the nonlinear autoencoding structure. Multiple IRMs can be flexibly cascaded to achieve the superposition of BCSs, thus, portraying diverse multimodal couplings. Furthermore, by balancing the stationarity requirement and reconstruction constraint, the designed compatibility learning algorithm induces the cascaded IRMs to capture and filter out nonstationarity, thus, obtaining stationary refined data. In this way, stationarity and nonstationarity with multimodal couplings can be fully separated. The validity of CDIS is illustrated through a simulated example and a real condensing system experimental rig.
引用
收藏
页码:8121 / 8131
页数:11
相关论文
共 50 条
  • [31] Video Representation Learning for Decoupled Deep Reinforcement Learning Applied to Autonomous Driving
    Mohammed, Shawan Taha
    Kastouri, Mohamed
    Niederfahrenhorst, Artur
    Ascheid, Gerd
    2023 IEEE/SICE INTERNATIONAL SYMPOSIUM ON SYSTEM INTEGRATION, SII, 2023,
  • [32] Multimodal pretraining for unsupervised protein representation learning
    Nguyen, Viet Thanh Duy
    Hy, Truong Son
    BIOLOGY METHODS & PROTOCOLS, 2024, 9 (01):
  • [33] Improved Multimodal Representation Learning with Skip Connections
    Zhang, Ning
    Cao, Yu
    Liu, Benyuan
    Luo, Yan
    PROCEEDINGS OF THE 2017 ACM MULTIMEDIA CONFERENCE (MM'17), 2017, : 654 - 662
  • [34] Variational Invariant Representation Learning for Multimodal Recommendation
    Yang, Wei
    Zhang, Haoran
    Zhang, Li
    PROCEEDINGS OF THE 2024 SIAM INTERNATIONAL CONFERENCE ON DATA MINING, SDM, 2024, : 752 - 760
  • [35] A deep semantic framework for multimodal representation learning
    Cheng Wang
    Haojin Yang
    Christoph Meinel
    Multimedia Tools and Applications, 2016, 75 : 9255 - 9276
  • [36] Survey of Research on Deep Multimodal Representation Learning
    Pan, Mengzhu
    Li, Qianmu
    Qiu, Tian
    Computer Engineering and Applications, 2024, 59 (02) : 48 - 64
  • [37] Unsupervised Natural Language Inference via Decoupled Multimodal Contrastive Learning
    Cui, Wanyun
    Zheng, Guangyu
    Wang, Wei
    PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 5511 - 5520
  • [38] Multimodal Representation Learning for Human Robot Interaction
    Sheppard, Eli
    Lohan, Katrin S.
    HRI'20: COMPANION OF THE 2020 ACM/IEEE INTERNATIONAL CONFERENCE ON HUMAN-ROBOT INTERACTION, 2020, : 445 - 446
  • [39] A deep semantic framework for multimodal representation learning
    Wang, Cheng
    Yang, Haojin
    Meinel, Christoph
    MULTIMEDIA TOOLS AND APPLICATIONS, 2016, 75 (15) : 9255 - 9276
  • [40] Adapt and explore: Multimodal mixup for representation learning
    Lin, Ronghao
    Hu, Haifeng
    Information Fusion, 2024, 105