PAC-Bayesian Contrastive Unsupervised Representation Learning

被引:0
|
作者
Nozawa, Kento [1 ,2 ]
Germain, Pascal [3 ]
Guedj, Benjamin [4 ,5 ]
机构
[1] Univ Tokyo, Tokyo, Japan
[2] RIKEN, Wako, Saitama, Japan
[3] Univ Laval, Quebec City, PQ, Canada
[4] Inria, Le Chesnay, France
[5] UCL, London, England
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Contrastive unsupervised representation learning (CURL) is the state-of-the-art technique to learn representations (as a set of features) from unlabelled data. While CURL has collected several empirical successes recently, theoretical understanding of its performance was still missing. In a recent work, Arora et al. (2019) provide the first generalisation bounds for CURL, relying on a Rademacher complexity. We extend their framework to the flexible PAC-Bayes setting, allowing us to deal with the non-iid setting. We present PAC-Bayesian generalisation bounds for CURL, which are then used to derive a new representation learning algorithm. Numerical experiments on real-life datasets illustrate that our algorithm achieves competitive accuracy, and yields nonvacuous generalisation bounds.
引用
下载
收藏
页码:21 / 30
页数:10
相关论文
共 50 条
  • [31] Improved PAC-Bayesian Bounds for Linear Regression
    Shalaeva, Vera
    Esfahani, Alireza Fakhrizadeh
    Germain, Pascal
    Petreczky, Mihaly
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 5660 - 5667
  • [32] A Theoretical Analysis of Contrastive Unsupervised Representation Learning
    Arora, Sanjeev
    Khandeparkar, Hrishikesh
    Khodak, Mikhail
    Plevrakis, Orestis
    Saunshi, Nikunj
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 97, 2019, 97
  • [33] Relative Contrastive Loss for Unsupervised Representation Learning
    Tang, Shixiang
    Zhu, Feng
    Bai, Lei
    Zhao, Rui
    Ouyang, Wanli
    COMPUTER VISION - ECCV 2022, PT XXVII, 2022, 13687 : 1 - 18
  • [34] PAC-Bayesian Bounds based on the Renyi Divergence
    Begin, Luc
    Germain, Pascal
    Laviolotte, Francois
    Roy, Jean-Francis
    ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 51, 2016, 51 : 435 - 444
  • [35] PAC-Bayesian theory for stochastic LTI systems
    Eringis, Deividas
    Leth, John
    Tan, Zheng-Hua
    Wisniewski, Rafal
    Esfahani, Alireza Fakhrizadeh
    Petreczky, Mihaly
    2021 60TH IEEE CONFERENCE ON DECISION AND CONTROL (CDC), 2021, : 6626 - 6633
  • [36] PAC-Bayesian Bound for the Conditional Value at Risk
    Mhammedi, Zakaria
    Guedj, Benjamin
    Williamson, Robert C.
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [37] Risk Bounds for the Majority Vote: From a PAC-Bayesian Analysis to a Learning Algorithm
    Germain, Pascal
    Lacasse, Alexandre
    Laviolette, Francois
    Marchand, Mario
    Roy, Jean-Francis
    JOURNAL OF MACHINE LEARNING RESEARCH, 2015, 16 : 787 - 860
  • [38] Simpler PAC-Bayesian bounds for hostile data
    Alquier, Pierre
    Guedj, Benjamin
    MACHINE LEARNING, 2018, 107 (05) : 887 - 902
  • [39] A PAC-Bayesian Generalization Bound for Equivariant Networks
    Behboodi, Arash
    Cesa, Gabriele
    Cohen, Taco
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [40] On the Importance of Gradient Norm in PAC-Bayesian Bounds
    Gat, Itai
    Adi, Yossi
    Schwing, Alexander
    Hazan, Tamir
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35, NEURIPS 2022, 2022,