Surprisal analysis

Last updated

Surprisal analysis is an information-theoretical analysis technique that integrates and applies principles of thermodynamics and maximal entropy. Surprisal analysis is capable of relating the underlying microscopic properties to the macroscopic bulk properties of a system. It has already been applied to a spectrum of disciplines including engineering, physics, chemistry and biomedical engineering. Recently, it has been extended to characterize the state of living cells, specifically monitoring and characterizing biological processes in real time using transcriptional data.

Contents

History

Surprisal analysis was formulated at the Hebrew University of Jerusalem as a joint effort between Raphael David Levine, Richard Barry Bernstein and Avinoam Ben-Shaul in 1972. Levine and colleagues had recognized a need to better understand the dynamics of non-equilibrium systems, particularly of small systems, that are not seemingly applicable to thermodynamic reasoning. [1] Alhassid and Levine first applied surprisal analysis in nuclear physics, to characterize the distribution of products in heavy ion reactions. Since its formulation, surprisal analysis has become a critical tool for the analysis of reaction dynamics and is an official IUPAC term. [2] *

A schematic of "Surprisal Analysis". Schematic of Surprisal Analysis.jpg
A schematic of “Surprisal Analysis".

Application

Maximum entropy methods are at the core of a new view of scientific inference, allowing analysis and interpretation of large and sometimes noisy data. Surprisal analysis extends principles of maximal entropy and of thermodynamics, where both equilibrium thermodynamics and statistical mechanics are assumed to be inferential processes. This enables surprisal analysis to be an effective method of information quantification and compaction and of providing an unbiased characterization of systems. Surprisal analysis is particularly useful to characterize and understand dynamics in small systems, where energy fluxes that are otherwise negligible in large systems heavily influence system behavior.

Foremost, surprisal analysis identifies the state of a system when it reaches its maximal entropy, or thermodynamic equilibrium. This is known as balance state of the system because once a system reaches its maximal entropy, it can no longer initiate or participate in spontaneous processes. Following the determination of the balanced state, surprisal analysis then characterizes all the states in which the system deviates away from the balance state. These deviations are caused by constraints; these constraints on the system prevent the system from reaching its maximal entropy. Surprisal analysis is applied to both identify and characterize these constraints. In terms of the constraints, the probability of an event is quantified by

.

Here is the probability of the event in the balanced state. It is usually called the “prior probability” because it is the probability of an event prior to any constraints. The surprisal itself is defined as

The surprisal equals the sum over the constraints and is a measure of the deviation from the balanced state. These deviations are ranked on the degree of deviation from the balance state and ordered on the most to least influential to the system. This ranking is provided through the use of Lagrange multipliers. The most important constraint and usually the constraint sufficient to characterize a system exhibit the largest Lagrange multiplier. The multiplier for constraint is denoted above as ; larger multipliers identify more influential constraints. The event variable is the value of the constraint for the event . Using the method of Lagrange multipliers [3] requires that the prior probability and the nature of the constraints be experimentally identified. A numerical algorithm for determining Lagrange multipliers has been introduced by Agmon et al. [4] Recently, singular value decomposition and principal component analysis of the surprisal was utilized to identify constraints on biological systems, extending surprisal analysis to better understanding biological dynamics as shown in the figure.

A schematic of "Surprisal Analysis". Surprisal Analysis of Gene Transcripts.jpg
A schematic of “Surprisal Analysis".

In physics

Surprisal (a term coined [5] in this context by Myron Tribus [6] ) was first introduced to better understand the specificity of energy release and selectivity of energy requirements of elementary chemical reactions. [1] This gave rise to a series of new experiments which demonstrated that in elementary reactions, the nascent products could be probed and that the energy is preferentially released and not statistically distributed. [1] Surprisal analysis was initially applied to characterize a small three molecule system that did not seemingly conform to principles of thermodynamics and a single dominant constraint was identified that was sufficient to describe the dynamic behavior of the three molecule system. Similar results were then observed in nuclear reactions, where differential states with varying energy partitioning are possible. Often chemical reactions require energy to overcome an activation barrier. Surprisal analysis is applicable to such applications as well. [7] Later, surprisal analysis was extended to mesoscopic systems, bulk systems [3] and to dynamical processes. [8]

In biology and biomedical sciences

Surprisal analysis was extended to better characterize and understand cellular processes, [9] see figure, biological phenomena and human disease with reference to personalized diagnostics. Surprisal analysis was first utilized to identify genes implicated in the balance state of cells in vitro; the genes mostly present in the balance state were genes directly responsible for the maintenance of cellular homeostasis. [10] Similarly, it has been used to discern two distinct phenotypes during the EMT of cancer cells. [11]

See also

Related Research Articles

In a chemical reaction, chemical equilibrium is the state in which both the reactants and products are present in concentrations which have no further tendency to change with time, so that there is no observable change in the properties of the system. This state results when the forward reaction proceeds at the same rate as the reverse reaction. The reaction rates of the forward and backward reactions are generally not zero, but they are equal. Thus, there are no net changes in the concentrations of the reactants and products. Such a state is known as dynamic equilibrium.

Entropy Property of a thermodynamic system

Entropy is a scientific concept as well as a measurable physical property that is most commonly associated with a state of disorder, randomness, or uncertainty. The term and the concept are used in diverse fields, from classical thermodynamics, where it was first recognized, to the microscopic description of nature in statistical physics, and to the principles of information theory. It has found far-ranging applications in chemistry and physics, in biological systems and their relation to life, in cosmology, economics, sociology, weather science, climate change, and information systems including the transmission of information in telecommunication.

The principle of maximum entropy states that the probability distribution which best represents the current state of knowledge about a system is the one with largest entropy, in the context of precisely stated prior data.

In classical statistical mechanics, the H-theorem, introduced by Ludwig Boltzmann in 1872, describes the tendency to decrease in the quantity H in a nearly-ideal gas of molecules. As this quantity H was meant to represent the entropy of thermodynamics, the H-theorem was an early demonstration of the power of statistical mechanics as it claimed to derive the second law of thermodynamics—a statement about fundamentally irreversible processes—from reversible microscopic mechanics. It is thought to prove the second law of thermodynamics, albeit under the assumption of low-entropy initial conditions.

Onsager reciprocal relations Relations between flows and forces, or gradients, in thermodynamic systems

In thermodynamics, the Onsager reciprocal relations express the equality of certain ratios between flows and forces in thermodynamic systems out of equilibrium, but where a notion of local equilibrium exists.

Non-equilibrium thermodynamics Branch of thermodynamics

Non-equilibrium thermodynamics is a branch of thermodynamics that deals with physical systems that are not in thermodynamic equilibrium but can be described in terms of macroscopic quantities that represent an extrapolation of the variables used to specify the system in thermodynamic equilibrium. Non-equilibrium thermodynamics is concerned with transport processes and with the rates of chemical reactions.

In information theory, the Rényi entropy generalizes the Hartley entropy, the Shannon entropy, the collision entropy and the min-entropy. Entropies quantify the diversity, uncertainty, or randomness of a system. The entropy is named after Alfréd Rényi, who looked for the most general definition of information measures that preserve additivity for independent events. In the context of fractal dimension estimation, the Rényi entropy forms the basis of the concept of generalized dimensions.

In statistics and information theory, a maximum entropy probability distribution has entropy that is at least as great as that of all other members of a specified class of probability distributions. According to the principle of maximum entropy, if nothing is known about a distribution except that it belongs to a certain class, then the distribution with the largest entropy should be chosen as the least-informative default. The motivation is twofold: first, maximizing entropy minimizes the amount of prior information built into the distribution; second, many physical systems tend to move towards maximal entropy configurations over time.

In physics, maximum entropy thermodynamics views equilibrium thermodynamics and statistical mechanics as inference processes. More specifically, MaxEnt applies inference techniques rooted in Shannon information theory, Bayesian probability, and the principle of maximum entropy. These techniques are relevant to any situation requiring prediction from incomplete or insufficient data. MaxEnt thermodynamics began with two papers by Edwin T. Jaynes published in the 1957 Physical Review.

The mathematical expressions for thermodynamic entropy in the statistical thermodynamics formulation established by Ludwig Boltzmann and J. Willard Gibbs in the 1870s are similar to the information entropy by Claude Shannon and Ralph Hartley, developed in the 1940s.

The concept entropy was first developed by German physicist Rudolf Clausius in the mid-nineteenth century as a thermodynamic property that predicts that certain spontaneous processes are irreversible or impossible. In statistical mechanics, entropy is formulated as a statistical property using probability theory. The statistical entropy perspective was introduced in 1870 by Austrian physicist Ludwig Boltzmann, who established a new field of physics that provided the descriptive linkage between the macroscopic observation of nature and the microscopic view based on the rigorous treatment of a large ensembles of microstates that constitute thermodynamic systems.

In computational chemistry, a constraint algorithm is a method for satisfying the Newtonian motion of a rigid body which consists of mass points. A restraint algorithm is used to ensure that the distance between mass points is maintained. The general steps involved are: (i) choose novel unconstrained coordinates, (ii) introduce explicit constraint forces, (iii) minimize constraint forces implicitly by the technique of Lagrange multipliers or projection methods.

Maximum entropy spectral estimation is a method of spectral density estimation. The goal is to improve the spectral quality based on the principle of maximum entropy. The method is based on choosing the spectrum which corresponds to the most random or the most unpredictable time series whose autocorrelation function agrees with the known values. This assumption, which corresponds to the concept of maximum entropy as used in both statistical mechanics and information theory, is maximally non-committal with regard to the unknown values of the autocorrelation function of the time series. It is simply the application of maximum entropy modeling to any type of spectrum and is used in all fields where data is presented in spectral form. The usefulness of the technique varies based on the source of the spectral data since it is dependent on the amount of assumed knowledge about the spectrum that can be applied to the model.

Biased random walk on a graph Structural analysis of a network

In network science, a biased random walk on a graph is a time path process in which an evolving variable jumps from its current state to one of various potential new states; unlike in a pure random walk, the probabilities of the potential new states are unequal.

A quantum heat engine is a device that generates power from the heat flow between hot and cold reservoirs. The operation mechanism of the engine can be described by the laws of quantum mechanics. The first realization of a quantum heat engine was pointed out by Scovil and Schulz-DuBois in 1959, showing the connection of efficiency of the Carnot engine and the 3-level maser. Quantum refrigerators share the structure of quantum heat engines with the purpose of pumping heat from a cold to a hot bath consuming power first suggested by Geusic, Schulz-DuBois, De Grasse and Scovil. When the power is supplied by a laser the process is termed optical pumping or laser cooling, suggested by Wineland and Hänsch. Surprisingly heat engines and refrigerators can operate up to the scale of a single particle thus justifying the need for a quantum theory termed quantum thermodynamics.

Quantum thermodynamics

Quantum thermodynamics is the study of the relations between two independent physical theories: thermodynamics and quantum mechanics. The two independent theories address the physical phenomena of light and matter. In 1905, Albert Einstein argued that the requirement of consistency between thermodynamics and electromagnetism leads to the conclusion that light is quantized obtaining the relation . This paper is the dawn of quantum theory. In a few decades quantum theory became established with an independent set of rules. Currently quantum thermodynamics addresses the emergence of thermodynamic laws from quantum mechanics. It differs from quantum statistical mechanics in the emphasis on dynamical processes out of equilibrium. In addition, there is a quest for the theory to be relevant for a single individual quantum system.

Coarse-grained modeling, coarse-grained models, aim at simulating the behaviour of complex systems using their coarse-grained (simplified) representation. Coarse-grained models are widely used for molecular modeling of biomolecules at various granularity levels.

Maximum-entropy random graph model

Maximum-entropy random graph models are random graph models used to study complex networks subject to the principle of maximum entropy under a set of structural constraints, which may be global, distributional, or local.

Soft configuration model Random graph model in applied mathematics

In applied mathematics, the soft configuration model (SCM) is a random graph model subject to the principle of maximum entropy under constraints on the expectation of the degree sequence of sampled graphs. Whereas the configuration model (CM) uniformly samples random graphs of a specific degree sequence, the SCM only retains the specified degree sequence on average over all network realizations; in this sense the SCM has very relaxed constraints relative to those of the CM. The SCM for graphs of size has a nonzero probability of sampling any graph of size , whereas the CM is restricted to only graphs having precisely the prescribed connectivity structure.

A set of networks that satisfies given structural characteristics can be treated as a network ensemble. Brought up by Ginestra Bianconi in 2007, the entropy of a network ensemble measures the level of the order or uncertainty of a network ensemble.

References

  1. 1 2 3 Levine, Raphael D. (2005). Molecular Reaction Dynamics. Cambridge University Press. ISBN   9780521842761.
  2. Agmon, N; Alhassid, Y; Levine, RD (1979). "An algorithm for finding the distribution of maximal entropy". Journal of Computational Physics. 30 (2): 250–258. Bibcode:1979JCoPh..30..250A. CiteSeerX   10.1.1.170.9363 . doi:10.1016/0021-9991(79)90102-5.
  3. 1 2 Levine, RD (1980). "An information theoretical approach to inversion problems". J. Phys. A. 13 (1): 91. Bibcode:1980JPhA...13...91L. doi:10.1088/0305-4470/13/1/011.
  4. Levine, RD; Bernstein, RB (1974). "Energy disposal and energy consumption in elementary chemical relations: The information theoretic approach". Acc. Chem. Res. 7 (12): 393–400. doi:10.1021/ar50084a001.
  5. Bernstein, R. B.; Levine, R. D. (1972). "Entropy and Chemical Change. I. Characterization of Product (and Reactant) Energy Distributions in Reactive Molecular Collisions: Information and Entropy Deficiency". The Journal of Chemical Physics. 57 (1): 434–449. Bibcode:1972JChPh..57..434B. doi:10.1063/1.1677983.
  6. Myron Tribus (1961) Thermodynamics and Thermostatics:An Introduction to Energy, Information and States of Matter, with Engineering Applications (D. Van Nostrand, 24 West 40 Street, New York 18, New York, U.S.A) Tribus, Myron (1961), pp. 64-66 borrow.
  7. Levine, RD (1978). "Information theory approach to molecular reaction dynamics". Annu. Rev. Phys. Chem. 29: 59–92. Bibcode:1978ARPC...29...59L. doi:10.1146/annurev.pc.29.100178.000423.
  8. Remacle, F; Levine, RD (1993). "Maximal entropy spectral fluctuations and the sampling of phase space". J. Chem. Phys. 99 (4): 2383–2395. Bibcode:1993JChPh..99.2383R. doi:10.1063/1.465253.
  9. Remacle, F; Kravchenko-Balasha, N; Levitzki, A; Levine, RD (June 1, 2010). "Information-theoretic analysis of phenotype changes in early stages of carcinogenesis". PNAS. 107 (22): 10324–29. Bibcode:2010PNAS..10710324R. doi: 10.1073/pnas.1005283107 . PMC   2890488 . PMID   20479229.
  10. Kravchenko-Balasha, Nataly; Levitzki, Alexander; Goldstein, Andrew; Rotter, Varda; Gross, A.; Remacle, F.; Levine, R. D. (March 20, 2012). "On a fundamental structure of gene networks in living cells". PNAS. 109 (12): 4702–4707. Bibcode:2012PNAS..109.4702K. doi: 10.1073/pnas.1200790109 . PMC   3311329 . PMID   22392990.
  11. Zadran, Sohila; Arumugam, Rameshkumar; Herschman, Harvey; Phelps, Michael; Levine, R. D. (August 3, 2014). "Surprisal analysis characterizes the free energy time course of cancer cells undergoing epithelial-to-mesenchymal transition". PNAS. 111 (36): 13235–13240. Bibcode:2014PNAS..11113235Z. doi: 10.1073/pnas.1414714111 . PMC   4246928 . PMID   25157127.