Neural mixture models with expectation-maximization for end-to-end deep clustering

被引:1
|
作者
Tissera, Dumindu [1 ,2 ]
Vithanage, Kasun [2 ]
Wijesinghe, Rukshan [1 ,2 ]
Xavier, Alex [2 ]
Jayasena, Sanath [2 ]
Fernando, Subha [2 ]
Rodrigo, Ranga [1 ,2 ]
机构
[1] Univeris Moratuwa, Dept Elect & Telecommun Engn, Moratuwa, Sri Lanka
[2] Univ Moratuwa, CodeGen QBITS Lab, Moratuwa, Sri Lanka
关键词
Deep clustering; Mixture models; Expectation; -maximization;
D O I
10.1016/j.neucom.2022.07.017
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Any clustering algorithm must synchronously learn to model the clusters and allocate data to those clus-ters in the absence of labels. Mixture model-based methods model clusters with pre-defined statistical distributions and allocate data to those clusters based on the cluster likelihoods. They iteratively refine those distribution parameters and member assignments following the Expectation-Maximization (EM) algorithm. However, the cluster representability of such hand-designed distributions that employ a lim-ited amount of parameters is not adequate for most real-world clustering tasks. In this paper, we realize mixture model-based clustering with a neural network where the final layer neurons, with the aid of an additional transformation, approximate cluster distribution outputs. The network parameters pose as the parameters of those distributions. The result is an elegant, much-generalized representation of clusters than a restricted mixture of hand-designed distributions. We train the network end-to-end via batch -wise EM iterations where the forward pass acts as the E-step and the backward pass acts as the M -step. In image clustering, the mixture-based EM objective can be used as the clustering objective along with existing representation learning methods. In particular, we show that when mixture-EM optimiza-tion is fused with consistency optimization, it improves the sole consistency optimization performance in clustering. Our trained networks outperform single-stage deep clustering methods that still depend on k -means, with unsupervised classification accuracy of 63.8% in STL10, 58% in CIFAR10, 25.9% in CIFAR100, and 98.9% in MNIST.(c) 2022 Elsevier B.V. All rights reserved.
引用
收藏
页码:249 / 262
页数:14
相关论文
共 50 条
  • [1] Learning Neural Models for End-to-End Clustering
    Meier, Benjamin Bruno
    Elezi, Ismail
    Amirian, Mohammadreza
    Duerr, Oliver
    Stadelmann, Thilo
    [J]. ARTIFICIAL NEURAL NETWORKS IN PATTERN RECOGNITION, ANNPR 2018, 2018, 11081 : 126 - 138
  • [2] ON THE BEHAVIOR OF THE EXPECTATION-MAXIMIZATION ALGORITHM FOR MIXTURE MODELS
    Barazandeh, Babak
    Razaviyayn, Meisam
    [J]. 2018 IEEE GLOBAL CONFERENCE ON SIGNAL AND INFORMATION PROCESSING (GLOBALSIP 2018), 2018, : 61 - 65
  • [3] Quantum Expectation-Maximization for Gaussian mixture models
    Kerenidis, Iordanis
    Luongo, Alessandro
    Prakash, Anupam
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 119, 2020, 119
  • [4] Neural Network based Explicit Mixture Models and Expectation-maximization based Learning
    Liu, Dong
    Vu, Minh Thanh
    Chatterjee, Saikat
    Rasmussen, Lars K.
    [J]. Proceedings of the International Joint Conference on Neural Networks, 2020,
  • [5] Neural Network based Explicit Mixture Models and Expectation-maximization based Learning
    Liu, Dong
    Minh Thanh Vu
    Chatterjee, Saikat
    Rasmussen, Lars K.
    [J]. 2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
  • [6] Expectation-Maximization for Adaptive Mixture Models in Graph Optimization
    Pfeifer, Tim
    Protzel, Peter
    [J]. 2019 INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2019, : 3151 - 3157
  • [7] END-TO-END DEEP MULTIMODAL CLUSTERING
    Zhang, Xianchao
    Mu, Jie
    Zong, Linlin
    Yang, Xiaochun
    [J]. 2020 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO (ICME), 2020,
  • [8] Deep expectation-maximization network for unsupervised image segmentation and clustering
    Pu, Yannan
    Sun, Jian
    Tang, Niansheng
    Xu, Zongben
    [J]. IMAGE AND VISION COMPUTING, 2023, 135
  • [9] An Expectation-Maximization algorithm for the Wishart mixture model: Application to movement clustering
    Hidot, Sullivan
    Saint-Jean, Christophe
    [J]. PATTERN RECOGNITION LETTERS, 2010, 31 (14) : 2318 - 2324
  • [10] IMPROVED GAUSSIAN MIXTURE MODEL WITH EXPECTATION-MAXIMIZATION FOR CLUSTERING OF REMOTE SENSING IMAGERY
    Neagoe, Victor-Emil
    Chirila-Berbentea, Vlad
    [J]. 2016 IEEE INTERNATIONAL GEOSCIENCE AND REMOTE SENSING SYMPOSIUM (IGARSS), 2016, : 3063 - 3065