A Unified Definition of Mutual Information with Applications in Machine Learning

被引:26
|
作者
Zeng, Guoping [1 ]
机构
[1] Elevate, Ft Worth, TX 76109 USA
关键词
D O I
10.1155/2015/201874
中图分类号
T [工业技术];
学科分类号
08 ;
摘要
There are various definitions of mutual information. Essentially, these definitions can be divided into two classes: (1) definitions with random variables and (2) definitions with ensembles. However, there are some mathematical flaws in these definitions. For instance, Class 1 definitions either neglect the probability spaces or assume the two random variables have the same probability space. Class 2 definitions redefine marginal probabilities from the joint probabilities. In fact, the marginal probabilities are given from the ensembles and should not be redefined from the joint probabilities. Both Class 1 and Class 2 definitions assume a joint distribution exists. Yet, they all ignore an important fact that the joint or the joint probability measure is not unique. In this paper, we first present a new unified definition of mutual information to cover all the various definitions and to fix their mathematical flaws. Our idea is to define the joint distribution of two random variables by taking the marginal probabilities into consideration. Next, we establish some properties of the newly defined mutual information. We then propose a method to calculate mutual information in machine learning. Finally, we apply our newly defined mutual information to credit scoring.
引用
收藏
页数:12
相关论文
共 50 条
  • [31] Mutual Information Divergence: A Unified Metric for Multimodal Generative Models
    Kim, Jin-Hwa
    Kim, Yunji
    Lee, Jiyoung
    Yoo, Kang Min
    Lee, Sang-Woo
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [32] Reply to "Comment on 'Multiparty quantum mutual information: An alternative definition' "
    Kumar, Asutosh
    PHYSICAL REVIEW A, 2023, 108 (06)
  • [33] Learning curves for mutual information maximization
    Urbanczik, R
    PHYSICAL REVIEW E, 2003, 68 (01): : 161061 - 161066
  • [34] Competitive learning by mutual information maximization
    Kamimura, R
    Kamimura, T
    IJCNN'01: INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, VOLS 1-4, PROCEEDINGS, 2001, : 926 - 931
  • [35] Learning Speaker Representations with Mutual Information
    Ravanelli, Mirco
    Bengio, Yoshua
    INTERSPEECH 2019, 2019, : 1153 - 1157
  • [36] Mutual Information Driven Federated Learning
    Uddin, Md Palash
    Xiang, Yong
    Lu, Xuequan
    Yearwood, John
    Gao, Longxiang
    IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2021, 32 (07) : 1526 - 1538
  • [37] An Improvised Machine Learning Model Based on Mutual Information Feature Selection Approach for Microbes Classification
    Dhindsa, Anaahat
    Bhatia, Sanjay
    Agrawal, Sunil
    Sohi, Balwinder Singh
    ENTROPY, 2021, 23 (02) : 1 - 16
  • [38] Quantifying Aleatoric and Epistemic Uncertainty in Machine Learning: Are Conditional Entropy and Mutual Information Appropriate Measures?
    Wimmer, Lisa
    Sale, Yusuf
    Hofman, Paul
    Bischl, Bernd
    Huellermeier, Eyke
    UNCERTAINTY IN ARTIFICIAL INTELLIGENCE, 2023, 216 : 2282 - 2292
  • [39] The Resolved Mutual Information Function as a Structural Fingerprint of Biomolecular Sequences for Interpretable Machine Learning Classifiers
    Bohnsack, Katrin Sophie
    Kaden, Marika
    Abel, Julia
    Saralajew, Sascha
    Villmann, Thomas
    ENTROPY, 2021, 23 (10)
  • [40] Towards a Unified Theory of Learning and Information
    Alabdulmohsin, Ibrahim
    ENTROPY, 2020, 22 (04)