Skip to main content
Article
Shannon Information Entropy in the Canonical Genetic Code
Journal of Theoretical Biology
  • Louis R. Nemzer, Nova Southeastern University
Publication Date
2-21-2017
Keywords
  • Genetic code,
  • RNA translation,
  • Information theory,
  • Shannon entropy,
  • Amino acids
Abstract

The Shannon entropy measures the expected information value of messages. As with thermodynamic entropy, the Shannon entropy is only defined within a system that identifies at the outset the collections of possible messages, analogous to microstates, that will be considered indistinguishable macrostates. This fundamental insight is applied here for the first time to amino acid alphabets, which group the twenty common amino acids into families based on chemical and physical similarities. To evaluate these schemas objectively, a novel quantitative method is introduced based the inherent redundancy in the canonical genetic code. Each alphabet is taken as a separate system that partitions the 64 possible RNA codons, the microstates, into families, the macrostates. By calculating the normalized mutual information, which measures the reduction in Shannon entropy, conveyed by single nucleotide messages, groupings that best leverage this aspect of fault tolerance in the code are identified. The relative importance of properties related to protein folding - like hydropathy and size - and function, including side-chain acidity, can also be estimated. This approach allows the quantification of the average information value of nucleotide positions, which can shed light on the coevolution of the canonical genetic code with the tRNA-protein translation mechanism.

Comments

©2016 Elsevier Ltd. All rights reserved.

DOI
10.1016/j.jtbi.2016.12.010
Citation Information
Louis R. Nemzer. "Shannon Information Entropy in the Canonical Genetic Code" Journal of Theoretical Biology Vol. 415 (2017) p. 158 - 170 ISSN: 0022-5193
Available at: http://works.bepress.com/lnemzer/34/