On Memorization in Probabilistic Deep Generative Models

被引:0
|
作者
van den Burg, Gerrit J. J. [2 ]
Williams, Christopher K. I. [1 ,2 ]
机构
[1] Univ Edinburgh, Edinburgh EH8 9YL, Midlothian, Scotland
[2] Alan Turing Inst, London, England
基金
英国工程与自然科学研究理事会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recent advances in deep generative models have led to impressive results in a variety of application domains. Motivated by the possibility that deep learning models might memorize part of the input data, there have been increased efforts to understand how memorization arises. In this work, we extend a recently proposed measure of memorization for supervised learning (Feldman, 2019) to the unsupervised density estimation problem and adapt it to be more computationally efficient. Next, we present a study that demonstrates how memorization can occur in probabilistic deep generative models such as variational autoencoders. This reveals that the form of memorization to which these models are susceptible differs fundamentally from mode collapse and overfitting. Furthermore, we show that the proposed memorization score measures a phenomenon that is not captured by commonly-used nearest neighbor tests. Finally, we discuss several strategies that can be used to limit memorization in practice. Our work thus provides a framework for understanding problematic memorization in probabilistic generative models.
引用
收藏
页数:13
相关论文
共 50 条
  • [1] Probabilistic forecasting using deep generative models
    Alessandro Fanfarillo
    Behrooz Roozitalab
    Weiming Hu
    Guido Cervone
    [J]. GeoInformatica, 2021, 25 : 127 - 147
  • [2] Probabilistic forecasting using deep generative models
    Fanfarillo, Alessandro
    Roozitalab, Behrooz
    Hu, Weiming
    Cervone, Guido
    [J]. GEOINFORMATICA, 2021, 25 (01) : 127 - 147
  • [3] Probabilistic Typology: Deep Generative Models of Vowel Inventories
    Cotterell, Ryan
    Eisner, Jason
    [J]. PROCEEDINGS OF THE 55TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2017), VOL 1, 2017, : 1182 - 1192
  • [4] Problems using deep generative models for probabilistic audio source separation
    Frank, Maurice
    Ilse, Maximilian
    [J]. NEURIPS WORKSHOPS, 2020, 2020, 137 : 53 - 59
  • [5] Robot Concept Acquisition Based on Interaction Between Probabilistic and Deep Generative Models
    Kuniyasu, Ryo
    Nakamura, Tomoaki
    Taniguchi, Tadahiro
    Nagai, Takayuki
    [J]. FRONTIERS IN COMPUTER SCIENCE, 2021, 3
  • [6] Toward Discriminating and Synthesizing Motion Traces Using Deep Probabilistic Generative Models
    Zhou, Fan
    Liu, Xin
    Zhang, Kunpeng
    Trajcevski, Goce
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2021, 32 (06) : 2401 - 2414
  • [7] Probabilistic harmonization and annotation of single-cell transcriptomics data with deep generative models
    Xu, Chenling
    Lopez, Romain
    Mehlman, Edouard
    Regier, Jeffrey
    Jordan, Michael, I
    Yosef, Nir
    [J]. MOLECULAR SYSTEMS BIOLOGY, 2021, 17 (01)
  • [8] Probabilistic generative transformer language models for generative design of molecules
    Wei, Lai
    Fu, Nihang
    Song, Yuqi
    Wang, Qian
    Hu, Jianjun
    [J]. JOURNAL OF CHEMINFORMATICS, 2023, 15 (01)
  • [9] Diversity in Deep Generative Models and Generative AI
    Turinici, Gabriel
    [J]. MACHINE LEARNING, OPTIMIZATION, AND DATA SCIENCE, LOD 2023, PT II, 2024, 14506 : 84 - 93
  • [10] Probabilistic generative transformer language models for generative design of molecules
    Lai Wei
    Nihang Fu
    Yuqi Song
    Qian Wang
    Jianjun Hu
    [J]. Journal of Cheminformatics, 15