The Shannon Information entropy of protein sequences

被引:130
|
作者
Strait, BJ [1 ]
Dewey, TG [1 ]
机构
[1] UNIV DENVER, DEPT CHEM, DENVER, CO 80208 USA
关键词
D O I
10.1016/S0006-3495(96)79210-X
中图分类号
Q6 [生物物理学];
学科分类号
071011 ;
摘要
A comprehensive data base is analyzed to determine the Shannon information content of a protein sequence. This information entropy is estimated by three methods: a k-tuplet analysis, a generalized Zipf analysis, and a ''Chou-Fasman gambler.'' The k-tuplet analysis is a ''letter'' analysis, based on conditional sequence probabilities. The generalized Zipf analysis demonstrates the statistical linguistic qualities of protein sequences and uses the ''word'' frequency to determine the Shannon entropy. The Zipf analysis and k-tuplet analysis give Shannon entropies of approximately 2.5 bits/amino acid. This entropy is much smaller than the value of 4.18 bits/amino acid obtained from the nonuniform composition of amino acids in proteins. The ''Chou-Fasman'' gambler is an algorithm based on the Chou-Fasman rules for protein structure. It uses both sequence and secondary structure information to guess at the number of-possible amino acids that could appropriately substitute into a sequence. As in the case for the English language, the gambler algorithm gives significantly lower entropies than the k-tuplet analysis. Using these entropies, the number of most probable protein sequences can be calculated. The number of most probable protein sequences is much less than the number of possible sequences but is still much larger than the number of sequences thought to have existed throughout evolution. Implications of these results for mutagenesis experiments are discussed.
引用
收藏
页码:148 / 155
页数:8
相关论文
共 50 条
  • [1] SHANNON-INFORMATION IS NOT ENTROPY
    SCHIFFER, M
    [J]. PHYSICS LETTERS A, 1991, 154 (7-8) : 361 - 365
  • [2] Physical information entropy and probability Shannon entropy
    R. Ascoli
    R. Urigu
    [J]. International Journal of Theoretical Physics, 1997, 36 : 1691 - 1716
  • [3] Physical information entropy and probability shannon entropy
    Ascoli, R
    Urigu, R
    [J]. INTERNATIONAL JOURNAL OF THEORETICAL PHYSICS, 1997, 36 (08) : 1691 - 1716
  • [4] Dynamical Shannon entropy and information Tsallis entropy in complex systems
    Yulmetyev, RM
    Emelyanova, NA
    Gafarov, FM
    [J]. PHYSICA A-STATISTICAL MECHANICS AND ITS APPLICATIONS, 2004, 341 : 649 - 676
  • [5] Shannon's Measure of Information and the Thermodynamic Entropy
    Ben-Naim, Arieh
    [J]. BAYESIAN INFERENCE AND MAXIMUM ENTROPY METHODS IN SCIENCE AND ENGINEERING, 2012, 1443 : 129 - 142
  • [6] Nonrelativistic Shannon information entropy for Kratzer potential
    Najafizade, S. A.
    Hassanabadi, H.
    Zarrinkamar, S.
    [J]. CHINESE PHYSICS B, 2016, 25 (04)
  • [7] Global Seismic Nowcasting With Shannon Information Entropy
    Rundle, John B.
    Giguere, Alexis
    Turcotte, Donald L.
    Crutchfield, James P.
    Donnellan, Andrea
    [J]. EARTH AND SPACE SCIENCE, 2019, 6 (01) : 191 - 197
  • [8] On Shannon-Jaynes entropy and fisher information
    Dimitrov, Vesselin I.
    [J]. BAYESIAN INFERENCE AND MAXIMUM ENTROPY METHODS IN SCIENCE AND ENGINEERING, 2007, 954 : 143 - 152
  • [9] Nonrelativistic Shannon information entropy for Kratzer potential
    Najafizade S A
    Hassanabadi H
    Zarrinkamar S
    [J]. Chinese Physics B, 2016, 25 (04) : 41 - 45
  • [10] Nonrelativistic Shannon information entropy for Killingbeck potential
    Najafizade, Seyede Amene
    Hassanabadi, Hassan
    Zarrinkamar, Saber
    [J]. CANADIAN JOURNAL OF PHYSICS, 2016, 94 (10) : 1085 - 1092