Development of a Fingerprint Reduction Approach for Bayesian Similarity Searching Based on Kullback-Leibler Divergence Analysis

被引:23
|
作者
Nisius, Britta [1 ]
Vogt, Martin [1 ]
Bajorath, Juergen [1 ]
机构
[1] Rhein Freidrich Wilhelms Univ Bonn, Dept Life Sci Informat, B IT, LIMES Program Unit Chem Biol & Med Chem, D-53113 Bonn, Germany
关键词
DIMENSIONAL DESCRIPTOR SPACES; ACTIVE COMPOUNDS; PERFORMANCE; MOLECULES; DATABASE; FUSION; 2D;
D O I
10.1021/ci900087y
中图分类号
R914 [药物化学];
学科分类号
100701 ;
摘要
The contribution of individual fingerprint bit positions to similarity search performance is systematically evaluated. A method is introduced to determine bit significance on the basis of Kullback-Leibler divergence analysis of bit distributions in active and database compounds. Bit divergence analysis and Bayesian compound screening share a common methodological foundation. Hence, given the significance ranking of all individual bit positions comprising a fingerprint, subsets of bits are evaluated in the context of Bayesian screening, and minimal fingerprint representations are determined that meet or exceed the search performance of unmodified fingerprints. For fingerprints of different design evaluated on many compound activity classes, we consistently find that subsets of fingerprint bit positions are responsible for search performance. In part, these subsets are very small and contain in some cases only a few fingerprint bit positions. Structural or pharmacophore patterns captured by preferred bit positions can often be directly associated with characteristic features of active compounds. In some cases, reduced fingerprint representations clearly exceed the search performance of the original fingerprints. Thus, fingerprint reduction likely represents a promising approach for practical applications.
引用
收藏
页码:1347 / 1358
页数:12
相关论文
共 50 条
  • [1] Bayesian similarity searching in high-dimensional descriptor spaces combined with Kullback-Leibler descriptor divergence analysis
    Vogt, Martin
    Bajorath, Jijrgen
    JOURNAL OF CHEMICAL INFORMATION AND MODELING, 2008, 48 (02) : 247 - 255
  • [2] Computation of Kullback-Leibler Divergence in Bayesian Networks
    Moral, Serafin
    Cano, Andres
    Gomez-Olmedo, Manuel
    ENTROPY, 2021, 23 (09)
  • [3] Kullback-Leibler divergence: A quantile approach
    Sankaran, P. G.
    Sunoj, S. M.
    Nair, N. Unnikrishnan
    STATISTICS & PROBABILITY LETTERS, 2016, 111 : 72 - 79
  • [4] Bayesian case influence analysis for GARCH models based on Kullback-Leibler divergence
    Hao, Hong-Xia
    Lin, Jin-Guan
    Wang, Hong-Xia
    Huang, Xing-Fang
    JOURNAL OF THE KOREAN STATISTICAL SOCIETY, 2016, 45 (04) : 595 - 609
  • [5] Assessing Similarity of Rating Distributions by Kullback-Leibler Divergence
    Corduas, Marcella
    CLASSIFICATION AND MULTIVARIATE ANALYSIS FOR COMPLEX DATA STRUCTURES, 2011, : 221 - 228
  • [6] Kullback-Leibler divergence for Bayesian nonparametric model checking
    Al-Labadi, Luai
    Patel, Viskakh
    Vakiloroayaei, Kasra
    Wan, Clement
    JOURNAL OF THE KOREAN STATISTICAL SOCIETY, 2021, 50 (01) : 272 - 289
  • [7] Modulation Classification Based on Kullback-Leibler Divergence
    Im, Chaewon
    Ahn, Seongjin
    Yoon, Dongweon
    15TH INTERNATIONAL CONFERENCE ON ADVANCED TRENDS IN RADIOELECTRONICS, TELECOMMUNICATIONS AND COMPUTER ENGINEERING (TCSET - 2020), 2020, : 373 - 376
  • [8] A similarity measure based on Kullback-Leibler divergence for collaborative filtering in sparse data
    Deng, Jiangzhou
    Wang, Yong
    Guo, Junpeng
    Deng, Yongheng
    Gao, Jerry
    Park, Younghee
    JOURNAL OF INFORMATION SCIENCE, 2019, 45 (05) : 656 - 675
  • [9] Approximate Bayesian Computation with Kullback-Leibler Divergence as Data Discrepancy
    Jiang, Bai
    Wu, Tung-Yu
    Wong, Wing Hung
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 84, 2018, 84