An Entropy-Based Model for Hierarchical Learning

被引:0
|
作者
Asadi, Amir R. [1 ]
机构
[1] Univ Cambridge, Ctr Math Sci, Stat Lab, Cambridge CB3 0WA, England
关键词
machine learning; neural network; chaining; information theory; scale-invariant distribution; curriculum learning; logarithmic binning; DISTRIBUTIONS;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Machine learning, the predominant approach in the field of artificial intelligence, enables computers to learn from data and experience. In the supervised learning framework, accurate and efficient learning of dependencies between data instances and their corresponding labels requires auxiliary information about the data distribution and the target function. This central concept aligns with the notion of regularization in statistical learning theory. Real-world datasets are often characterized by multiscale data instance distributions and well-behaved, smooth target functions. Scale-invariant probability distributions, such as power-law distributions, provide notable examples of multiscale data instance distributions in various contexts. This paper introduces a hierarchical learning model that leverages such a multiscale data structure with a multiscale entropy-based training procedure and explores its statistical and computational advantages. The hierarchical learning model is inspired by the logical progression in human learning from easy to complex tasks and features interpretable levels. In this model, the logarithm of any data instance's norm can be construed as the data instance's complexity, and the allocation of computational resources is tailored to this complexity, resulting in benefits such as increased inference speed. Furthermore, our multiscale analysis of the statistical risk yields stronger guarantees compared to conventional uniform convergence bounds.
引用
收藏
页码:1 / 45
页数:45
相关论文
共 50 条
  • [1] Entropy-based learning of sensing matrices
    Parthasarathy, Gayatri
    Abhilash, G.
    [J]. IET SIGNAL PROCESSING, 2019, 13 (07) : 650 - 660
  • [2] Entropy-based transform learning algorithms
    Parthasarathy, Gayatri
    Abhilash, G.
    [J]. IET SIGNAL PROCESSING, 2018, 12 (04) : 439 - 446
  • [3] Graph Entropy-Based Learning Analytics
    Al-Zawqari, Ali
    Vandersteen, Gerd
    [J]. ARTIFICIAL INTELLIGENCE IN EDUCATION: POSTERS AND LATE BREAKING RESULTS, WORKSHOPS AND TUTORIALS, INDUSTRY AND INNOVATION TRACKS, PRACTITIONERS AND DOCTORAL CONSORTIUM, PT II, 2022, 13356 : 16 - 21
  • [4] Entropy-Based Feature Extraction Model for Fundus Images with Deep Learning Model
    Gadde, Sai Sudha
    Kiran, K. V. D.
    [J]. INTERNATIONAL JOURNAL OF IMAGE AND GRAPHICS, 2022,
  • [5] Unsupervised non-hierarchical entropy-based clustering
    Jardino, M
    [J]. DATA ANALYSIS, CLASSIFICATION, AND RELATED METHODS, 2000, : 29 - 34
  • [6] AN ENTROPY-BASED MODAL SPLIT MODEL
    JORNSTEN, KO
    LUNDGREN, JT
    [J]. TRANSPORTATION RESEARCH PART B-METHODOLOGICAL, 1989, 23 (05) : 345 - 359
  • [7] Learning Sparse Codes with Entropy-Based ELBOs
    Velychko, Dmytro
    Damm, Simon
    Fischer, Asja
    Luecke, Joerg
    [J]. INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 238, 2024, 238
  • [8] Entropy-based closure for probabilistic learning on manifolds
    Soize, C.
    Ghanem, R.
    Safta, C.
    Huan, X.
    Vane, Z. P.
    Oefelein, J.
    Lacaze, G.
    Najm, H. N.
    Tang, Q.
    Chen, X.
    [J]. JOURNAL OF COMPUTATIONAL PHYSICS, 2019, 388 : 518 - 533
  • [9] Entropy-Based Active Learning for Object Recognition
    Holub, Alex
    Perona, Pietro
    Burl, Michael C.
    [J]. 2008 IEEE COMPUTER SOCIETY CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS, VOLS 1-3, 2008, : 885 - +
  • [10] An Improved Entropy-Based Multiple Kernel Learning
    Hino, Hideitsu
    Ogawa, Tetsuji
    [J]. 2012 21ST INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR 2012), 2012, : 1189 - 1192