A Unified Definition of Mutual Information with Applications in Machine Learning

被引:26
|
作者
Zeng, Guoping [1 ]
机构
[1] Elevate, Ft Worth, TX 76109 USA
关键词
D O I
10.1155/2015/201874
中图分类号
T [工业技术];
学科分类号
08 ;
摘要
There are various definitions of mutual information. Essentially, these definitions can be divided into two classes: (1) definitions with random variables and (2) definitions with ensembles. However, there are some mathematical flaws in these definitions. For instance, Class 1 definitions either neglect the probability spaces or assume the two random variables have the same probability space. Class 2 definitions redefine marginal probabilities from the joint probabilities. In fact, the marginal probabilities are given from the ensembles and should not be redefined from the joint probabilities. Both Class 1 and Class 2 definitions assume a joint distribution exists. Yet, they all ignore an important fact that the joint or the joint probability measure is not unique. In this paper, we first present a new unified definition of mutual information to cover all the various definitions and to fix their mathematical flaws. Our idea is to define the joint distribution of two random variables by taking the marginal probabilities into consideration. Next, we establish some properties of the newly defined mutual information. We then propose a method to calculate mutual information in machine learning. Finally, we apply our newly defined mutual information to credit scoring.
引用
收藏
页数:12
相关论文
共 50 条
  • [1] Machine Learning with Squared-Loss Mutual Information
    Sugiyama, Masashi
    ENTROPY, 2013, 15 (01) : 80 - 112
  • [2] Mutual information scaling for tensor network machine learning
    Convy, Ian
    Huggins, William
    Liao, Haoran
    Birgitta Whaley, K.
    MACHINE LEARNING-SCIENCE AND TECHNOLOGY, 2022, 3 (01):
  • [3] Ray as the Unified Compute Substrate for Machine Learning Applications
    Zhang, Zhe
    2021 IEEE INTERNATIONAL PARALLEL AND DISTRIBUTED PROCESSING SYMPOSIUM WORKSHOPS (IPDPSW), 2021, : 912 - 912
  • [4] Applications of machine learning in information retrieval
    Cunningham, SJ
    Witten, IH
    Littin, J
    ANNUAL REVIEW OF INFORMATION SCIENCE AND TECHNOLOGY, 1999, 34 : 341 - 384
  • [5] DEFINITION OF CONDITIONAL MUTUAL INFORMATION FOR ARBITRARY ENSEMBLES
    WYNER, AD
    INFORMATION AND CONTROL, 1978, 38 (01): : 51 - 59
  • [6] Multiparty quantum mutual information: An alternative definition
    Kumar, Asutosh
    PHYSICAL REVIEW A, 2017, 96 (01)
  • [7] Answer Extraction for Definition Questions Using Information Gain and Machine Learning
    Instituto Nacional de Astrofísica Óptica y Electrónica, Universidad de la Sierra Juárez, Mexico
    不详
    72840, Mexico
    IFIP Advances in Information and Communication Technology, 2008, (141-150)
  • [8] Answer extraction for definition questions using information gain and machine learning
    Martinez-Gil, Carmen
    Lopez-Lopez, A.
    ARTIFICIAL INTELLIGENCE IN THEORY AND PRACTICE II, 2008, 276 : 141 - 150
  • [9] A unified mathematical definition of classical information retrieval
    Dominich, S
    JOURNAL OF THE AMERICAN SOCIETY FOR INFORMATION SCIENCE, 2000, 51 (07): : 614 - 624
  • [10] MUTUAL INFORMATION PRINCIPLE AND APPLICATIONS
    TZANNES, NS
    NOONAN, JP
    INFORMATION AND CONTROL, 1973, 22 (01): : 1 - 12