Information fluctuation complexity

Last updated

Information fluctuation complexity is an information-theoretic quantity defined as the fluctuation of information about entropy. It is derivable from fluctuations in the predominance of order and chaos in a dynamic system and has been used as a measure of complexity in many diverse fields. It was introduced in a 1993 paper by Bates and Shepard. [1]

Contents

Definition

The information fluctuation complexity of a discrete dynamic system is a function of the probability distribution of its states when it is subject to random external input data. The purpose of driving the system with a rich information source such as a random number generator or a white noise signal is to probe the internal dynamics of the system much the same as a frequency-rich impulse is used in signal processing.

If a system has possible states and the state probabilitiesare known, then its information entropy is

where is the information content of state .

The information fluctuation complexity of the system is defined as the standard deviation or fluctuation of about its mean :

or

The fluctuation of state information is zero in a maximally disordered system with all ; the system simply mimics its random inputs. is also zero when the system is perfectly ordered with just one fixed state , regardless of inputs. is non-zero between these two extremes with a mixture of both higher-probability states and lower-probability states populating state space.

Fluctuation of information allows for memory and computation

As a complex dynamic system evolves in time, how it transitions between states depends on external stimuli in an irregular way. At times it may be more sensitive to external stimuli (unstable) and at other times less sensitive (stable). If a particular state has several possible next-states, external information determines which one will be next and the system gains that information by following a particular trajectory in state space. But if several different states all lead to the same next-state, then upon entering the next-state the system loses information about which state preceded it. Thus, a complex system exhibits alternating information gain and loss as it evolves in time. The alternation or fluctuation of information is equivalent to remembering and forgetting — temporary information storage or memory — an essential feature of non-trivial computation.

The gain or loss of information associated with transitions between states can be related to state information. The net information gain of a transition from state to state is the information gained when leaving state less the information lost when entering state :

Here is the forward conditional probability that if the present state is then the next state is and is the reverse conditional probability that if the present state is then the previous state was . The conditional probabilities are related to the transition probability, the probability that a transition from state to state occurs, by:

Eliminating the conditional probabilities:

Therefore the net information gained by the system as a result of the transition depends only on the increase in state information from the initial to the final state. It can be shown that this is true even for multiple consecutive transitions. [1]

is reminiscent of the relation between force and potential energy. is like potential and is like force in . External information “pushes” a system “uphill” to a state of higher information potential to accomplish memory storage, much like pushing a mass uphill to a state of higher gravitational potential stores energy. The amount of energy storage depends only on the final height, not the path up the hill. Likewise, the amount of information storage does not depend on the transition path between two states in state space. Once a system reaches a rare state with high information potential, it may "fall" to a more common state, losing the previously stored information.

It may be useful to compute the standard deviation of about its mean (which is zero), namely the fluctuation of net information gain, [1] but takes into account multi-transition memory loops in state space and therefore should be a better indicator of the computational power of a system. Moreover, is easier to calculate because there can be many more transitions than states.

Chaos and order

A dynamic system that is sensitive to external information (unstable) exhibits chaotic behavior whereas one that is insensitive to external information (stable) exhibits orderly behavior. A complex system exhibits both behaviors, fluctuating between them in dynamic balance when subject to a rich information source. The degree of fluctuation is quantified by ; it captures the alternation in the predominance of chaos and order in a complex system as it evolves in time.

Example: rule 110 variant of the elementary cellular automaton

The rule 110 variant of the elementary cellular automaton has been proven to be capable of universal computation. The proof is based on the existence and interactions of cohesive and self-perpetuating cell patterns known as gliders or spaceships, emergent phenomena, that imply the capability of groups of automaton cells to remember that a glider is passing through them. It is therefore to be expected that there will be memory loops in state space resulting from alternations of information gain and loss, instability and stability, chaos and order.

Consider a 3-cell group of adjacent automaton cells that obey rule 110: end-center-end. The next state of the center cell depends on the present state of itself and the end cells as specified by the rule:

Elementary cellular automaton rule 110
3-cell group1-1-11-1-01-0-11-0-00-1-10-1-00-0-10-0-0
next center cell01101110

To calculate the information fluctuation complexity of this system, attach a driver cell to each end of the 3-cell group to provide a random external stimulus like so, driver→end-center-end←driver, such that the rule can be applied to the two end cells. Next determine what the next state is for each possible present state and for each possible combination of driver cell contents, to determine the forward conditional probabilities.

The state diagram of this system is depicted below, with circles representing the states and arrows representing transitions between states. The eight states of this system, 1-1-1 to 0-0-0 are labeled with the octal equivalent of the 3-bit contents of the 3-cell group: 7 to 0. The transition arrows are labeled with forward conditional probabilities. Notice that there is variability in the divergence and convergence of arrows corresponding to a variability in chaos and order, sensitivity and insensitivity, gain and loss of external information from the driver cells.

The 3-cell state diagram for the rule 110 elementary cellular automaton showing forward conditional transition probabilities with random stimulation. State diagram for rule 110 with transition probabilities.png
The 3-cell state diagram for the rule 110 elementary cellular automaton showing forward conditional transition probabilities with random stimulation.

The forward conditional probabilities are determined by the proportion of possible driver cell contents that drive a particular transition. For example, for the four possible combinations of two driver cell contents, state 7 leads to states 5, 4, 1 and 0 so , , , and are each ¼ or 25%. Likewise, state 0 leads to states 0, 1, 0 and 1 so and are each ½ or 50%. And so forth.

The state probabilities are related by

and

These linear algebraic equations can be solved manually or with the aid of a computer program for the state probabilities, with the following results:

Rule 110 state probabilities for 3-cell automaton with random stimulation
p0p1p2p3p4p5p6p7
2/172/171/345/342/172/172/174/17

Information entropy and complexity can then be calculated from the state probabilities:

Note that the maximum possible entropy for eight states is which would be the case if all eight states were equally likely with probabilities of ⅛ (randomness). Thus rule 110 has a relatively high entropy or state utilization at 2.86 bits. But this does not preclude a substantial fluctuation of state information about entropy and thus a substantial value of complexity. Whereas, maximum entropy would preclude complexity.

An alternative method can be used to obtain the state probabilities when the analytical method used above is unfeasible. Simply drive the system at its inputs (the driver cells) with a random source for many generations and observe the state probabilities empirically. When this is done via computer simulation for 10 million generations the results are as follows: [2]

Information variables for the rule 110 elementary cellular automaton
number of cells345678910111213
(bits)2.863.814.735.666.567.478.349.2510.0910.9711.78
(bits)0.560.650.720.730.790.810.890.901.001.011.15
0.200.170.150.130.120.110.110.100.100.090.10

Since both and increase with system size, their dimensionless ratio , the relative information fluctuation complexity, is included to better compare systems of different sizes. Notice that the empirical and analytical results agree for the 3-cell automaton.

In the paper by Bates and Shepard, [1] is computed for all elementary cellular automaton rules and it was observed that the ones that exhibit slow-moving gliders and possibly stationary objects, as rule 110 does, are highly correlated with large values of . can therefore be used as a filter to select candidate rules for universal computation, which is tedious to prove.


Applications

Although the derivation of the information fluctuation complexity formula is based on information fluctuations in a dynamic system, the formula depends only on state probabilities and so is also applicable to any probability distribution, including those derived from static images or text.

Over the years the original paper [1] has been referred to by researchers in many diverse fields: complexity theory, [3] complex systems science, [4] complex networks, [5] chaotic dynamics, [6] many-body localization entanglement, [7] environmental engineering, [8] ecological complexity, [9] ecological time-series analysis, [10] ecosystem sustainability, [11] air [12] and water [13] pollution, hydrological wavelet analysis, [14] soil water flow, [15] soil moisture, [16] headwater runoff, [17] groundwater depth, [18] air traffic control, [19] flow patterns [20] and flood events, [21] topology, [22] economics, [23] market forecasting of metal [24] and electricity [25] prices, health informatics, [26] human cognition, [27] human gait kinematics, [28] neurology, [29] EEG analysis, [30] education, [31] investing, [32] artificial life [33] and aesthetics. [34]

Related Research Articles

In physics, the cross section is a measure of the probability that a specific process will take place when some kind of radiant excitation intersects a localized phenomenon. For example, the Rutherford cross-section is a measure of probability that an alpha particle will be deflected by a given angle during an interaction with an atomic nucleus. Cross section is typically denoted σ (sigma) and is expressed in units of area, more specifically in barns. In a way, it can be thought of as the size of the object that the excitation must hit in order for the process to occur, but more exactly, it is a parameter of a stochastic process.

<span class="mw-page-title-main">Entropy (information theory)</span> Expected amount of information needed to specify the output of a stochastic data source

In information theory, the entropy of a random variable is the average level of "information", "surprise", or "uncertainty" inherent to the variable's possible outcomes. Given a discrete random variable , which takes values in the alphabet and is distributed according to :

In quantum mechanics, a density matrix is a matrix that describes the quantum state of a physical system. It allows for the calculation of the probabilities of the outcomes of any measurement performed upon this system, using the Born rule. It is a generalization of the more usual state vectors or wavefunctions: while those can only represent pure states, density matrices can also represent mixed states. Mixed states arise in quantum mechanics in two different situations:

  1. when the preparation of the system is not fully known, and thus one must deal with a statistical ensemble of possible preparations, and
  2. when one wants to describe a physical system that is entangled with another, without describing their combined state; this case is typical for a system interacting with some environment.
<span class="mw-page-title-main">Log-normal distribution</span> Probability distribution

In probability theory, a log-normal (or lognormal) distribution is a continuous probability distribution of a random variable whose logarithm is normally distributed. Thus, if the random variable X is log-normally distributed, then Y = ln(X) has a normal distribution. Equivalently, if Y has a normal distribution, then the exponential function of Y, X = exp(Y), has a log-normal distribution. A random variable which is log-normally distributed takes only positive real values. It is a convenient and useful model for measurements in exact and engineering sciences, as well as medicine, economics and other topics (e.g., energies, concentrations, lengths, prices of financial instruments, and other metrics).

The fluctuation theorem (FT), which originated from statistical mechanics, deals with the relative probability that the entropy of a system which is currently away from thermodynamic equilibrium will increase or decrease over a given amount of time. While the second law of thermodynamics predicts that the entropy of an isolated system should tend to increase until it reaches equilibrium, it became apparent after the discovery of statistical mechanics that the second law is only a statistical one, suggesting that there should always be some nonzero probability that the entropy of an isolated system might spontaneously decrease; the fluctuation theorem precisely quantifies this probability.

<span class="mw-page-title-main">Mutual information</span> Measure of dependence between two variables

In probability theory and information theory, the mutual information (MI) of two random variables is a measure of the mutual dependence between the two variables. More specifically, it quantifies the "amount of information" obtained about one random variable by observing the other random variable. The concept of mutual information is intimately linked to that of entropy of a random variable, a fundamental notion in information theory that quantifies the expected "amount of information" held in a random variable.

In information geometry, the Fisher information metric is a particular Riemannian metric which can be defined on a smooth statistical manifold, i.e., a smooth manifold whose points are probability measures defined on a common probability space. It can be used to calculate the informational difference between measurements.

In quantum physics, a measurement is the testing or manipulation of a physical system to yield a numerical result. A fundamental feature of quantum theory is that the predictions it makes are probabilistic. The procedure for finding a probability involves combining a quantum state, which mathematically describes a quantum system, with a mathematical representation of the measurement to be performed on that system. The formula for this calculation is known as the Born rule. For example, a quantum particle like an electron can be described by a quantum state that associates to each point in space a complex number called a probability amplitude. Applying the Born rule to these amplitudes gives the probabilities that the electron will be found in one region or another when an experiment is performed to locate it. This is the best the theory can do; it cannot say for certain where the electron will be found. The same quantum state can also be used to make a prediction of how the electron will be moving, if an experiment is performed to measure its momentum instead of its position. The uncertainty principle implies that, whatever the quantum state, the range of predictions for the electron's position and the range of predictions for its momentum cannot both be narrow. Some quantum states imply a near-certain prediction of the result of a position measurement, but the result of a momentum measurement will be highly unpredictable, and vice versa. Furthermore, the fact that nature violates the statistical conditions known as Bell inequalities indicates that the unpredictability of quantum measurement results cannot be explained away as due to ignorance about "local hidden variables" within quantum systems.

In mathematical statistics, the Kullback–Leibler divergence, denoted , is a type of statistical distance: a measure of how one probability distribution P is different from a second, reference probability distribution Q. A simple interpretation of the KL divergence of P from Q is the expected excess surprise from using Q as a model when the actual distribution is P. While it is a measure of how different two distributions are, and in some sense is thus a "distance", it is not actually a metric, which is the most familiar and formal type of distance. In particular, it is not symmetric in the two distributions, and does not satisfy the triangle inequality. Instead, in terms of information geometry, it is a type of divergence, a generalization of squared distance, and for certain classes of distributions, it satisfies a generalized Pythagorean theorem.

In information theory, the Rényi entropy is a quantity that generalizes various notions of entropy, including Hartley entropy, Shannon entropy, collision entropy, and min-entropy. The Rényi entropy is named after Alfréd Rényi, who looked for the most general way to quantify information while preserving additivity for independent events. In the context of fractal dimension estimation, the Rényi entropy forms the basis of the concept of generalized dimensions.

<span class="mw-page-title-main">Satellite galaxy</span> Galaxy that orbits a larger galaxy due to gravitational attraction

A satellite galaxy is a smaller companion galaxy that travels on bound orbits within the gravitational potential of a more massive and luminous host galaxy. Satellite galaxies and their constituents are bound to their host galaxy, in the same way that planets within our own solar system are gravitationally bound to the Sun. While most satellite galaxies are dwarf galaxies, satellite galaxies of large galaxy clusters can be much more massive. The Milky Way is orbited by about fifty satellite galaxies, the largest of which is the Large Magellanic Cloud.

In quantum field theory and statistical mechanics, the Mermin–Wagner theorem states that continuous symmetries cannot be spontaneously broken at finite temperature in systems with sufficiently short-range interactions in dimensions d ≤ 2. Intuitively, this means that long-range fluctuations can be created with little energy cost, and since they increase the entropy, they are favored.

In mathematics, the topological entropy of a topological dynamical system is a nonnegative extended real number that is a measure of the complexity of the system. Topological entropy was first introduced in 1965 by Adler, Konheim and McAndrew. Their definition was modelled after the definition of the Kolmogorov–Sinai, or metric entropy. Later, Dinaburg and Rufus Bowen gave a different, weaker definition reminiscent of the Hausdorff dimension. The second definition clarified the meaning of the topological entropy: for a system given by an iterated function, the topological entropy represents the exponential growth rate of the number of distinguishable orbits of the iterates. An important variational principle relates the notions of topological and measure-theoretic entropy.

In quantum information theory, quantum relative entropy is a measure of distinguishability between two quantum states. It is the quantum mechanical analog of relative entropy.

In mathematics and economics, transportation theory or transport theory is a name given to the study of optimal transportation and allocation of resources. The problem was formalized by the French mathematician Gaspard Monge in 1781.

In quantum information theory, strong subadditivity of quantum entropy (SSA) is the relation among the von Neumann entropies of various quantum subsystems of a larger quantum system consisting of three subsystems. It is a basic theorem in modern quantum information theory. It was conjectured by D. W. Robinson and D. Ruelle in 1966 and O. E. Lanford III and D. W. Robinson in 1968 and proved in 1973 by E.H. Lieb and M.B. Ruskai, building on results obtained by Lieb in his proof of the Wigner-Yanase-Dyson conjecture.

The min-entropy, in information theory, is the smallest of the Rényi family of entropies, corresponding to the most conservative way of measuring the unpredictability of a set of outcomes, as the negative logarithm of the probability of the most likely outcome. The various Rényi entropies are all equal for a uniform distribution, but measure the unpredictability of a nonuniform distribution in different ways. The min-entropy is never greater than the ordinary or Shannon entropy and that in turn is never greater than the Hartley or max-entropy, defined as the logarithm of the number of outcomes with nonzero probability.

Generalized relative entropy is a measure of dissimilarity between two quantum states. It is a "one-shot" analogue of quantum relative entropy and shares many properties of the latter quantity.

Many-body localization (MBL) is a dynamical phenomenon which leads to the breakdown of equilibrium statistical mechanics in isolated many-body systems. Such systems never reach local thermal equilibrium, and retain local memory of their initial conditions for infinite times. One can still define a notion of phase structure in these out-of-equilibrium systems. Strikingly, MBL can even enable new kinds of exotic orders that are disallowed in thermal equilibrium – a phenomenon that goes by the name of localization-protected quantum order (LPQO) or eigenstate order.

A set of networks that satisfies given structural characteristics can be treated as a network ensemble. Brought up by Ginestra Bianconi in 2007, the entropy of a network ensemble measures the level of the order or uncertainty of a network ensemble.

References

  1. 1 2 3 4 5 Bates, John E.; Shepard, Harvey K. (1993-01-18). "Measuring complexity using information fluctuation". Physics Letters A. 172 (6): 416–425. Bibcode:1993PhLA..172..416B. doi:10.1016/0375-9601(93)90232-O. ISSN   0375-9601.
  2. Bates, John E. (2020-03-30). "Measuring complexity using information fluctuation: a tutorial". Research Gate.
  3. Atmanspacher, Harald (September 1997). "Cartesian cut, Heisenberg cut, and the concept of complexity". World Futures. 49 (3–4): 333–355. doi:10.1080/02604027.1997.9972639. ISSN   0260-4027.
  4. Shalizi, Cosma Rohilla (2006), Deisboeck, Thomas S.; Kresh, J. Yasha (eds.), "Methods and Techniques of Complex Systems Science: An Overview", Complex Systems Science in Biomedicine, Topics in Biomedical Engineering International Book Series, Springer US, pp. 33–114, arXiv: nlin/0307015 , doi:10.1007/978-0-387-33532-2_2, ISBN   978-0-387-33532-2, S2CID   11972113
  5. Huang, Min; Sun, Zhongkui; Donner, Reik V.; Zhang, Jie; Gua, Shuguang; Zou, Yong (2021-03-09). "Characterizing dynamical transitions by statistical complexity measures based on ordinal pattern transition networks". Chaos: An Interdisciplinary Journal of Nonlinear Science . 31 (3): 033127. doi:10.1063/5.0038876. PMID   33810737. S2CID   232771788.
  6. Wackerbauer, Renate (1995-11-01). "Noise-induced stabilization of the Lorenz system". Physical Review E. 52 (5): 4745–4749. Bibcode:1995PhRvE..52.4745W. doi:10.1103/PhysRevE.52.4745. PMID   9963970.
  7. Hamilton, Gregory A.; Clark, Bryan K. (2023-02-14). "Quantifying unitary flow efficiency and entanglement for many-body localization". Physical Review B. 107 (6): 064203. arXiv: 2110.10148 . doi:10.1103/PhysRevB.107.064203. S2CID   239024666.
  8. Singh, Vijay P. (2013-01-10). Entropy Theory and its Application in Environmental and Water Engineering. John Wiley & Sons. ISBN   978-1-118-42860-3.
  9. Parrott, Lael (2010-11-01). "Measuring ecological complexity". Ecological Indicators. 10 (6): 1069–1076. doi:10.1016/j.ecolind.2010.03.014. ISSN   1470-160X.
  10. Lange, Holger (2006), "Time-series Analysis in Ecology", eLS, American Cancer Society, doi:10.1038/npg.els.0003276, ISBN   978-0-470-01590-2
  11. Wang, Chaojun; Zhao, Hongrui (2019-04-18). "Analysis of remote sensing time-series data to foster ecosystem sustainability: use of temporal information entropy". International Journal of Remote Sensing. 40 (8): 2880–2894. Bibcode:2019IJRS...40.2880W. doi:10.1080/01431161.2018.1533661. ISSN   0143-1161. S2CID   135003743.
  12. Klemm, Otto; Lange, Holger (1999-12-01). "Trends of air pollution in the Fichtelgebirge Mountains, Bavaria". Environmental Science and Pollution Research. 6 (4): 193–199. doi:10.1007/BF02987325. ISSN   1614-7499. PMID   19005662. S2CID   35043.
  13. Wang, Kang; Lin, Zhongbing (2018). "Characterization of the nonpoint source pollution into river at different spatial scales". Water and Environment Journal. 32 (3): 453–465. doi:10.1111/wej.12345. ISSN   1747-6593. S2CID   115667734.
  14. Labat, David (2005-11-25). "Recent advances in wavelet analyses: Part 1. A review of concepts". Journal of Hydrology. 314 (1): 275–288. Bibcode:2005JHyd..314..275L. doi:10.1016/j.jhydrol.2005.04.003. ISSN   0022-1694.
  15. Pachepsky, Yakov; Guber, Andrey; Jacques, Diederik; Simunek, Jiri; Van Genuchten, Marthinus Th.; Nicholson, Thomas; Cady, Ralph (2006-10-01). "Information content and complexity of simulated soil water fluxes". Geoderma. Fractal Geometry Applied to Soil and Related Hierarchical Systems - Fractals, Complexity and Heterogeneity. 134 (3): 253–266. Bibcode:2006Geode.134..253P. doi:10.1016/j.geoderma.2006.03.003. ISSN   0016-7061.
  16. Kumar, Sujay V.; Dirmeyer, Paul A.; Peters-Lidard, Christa D.; Bindlish, Rajat; Bolten, John (2018-01-01). "Information theoretic evaluation of satellite soil moisture retrievals". Remote Sensing of Environment. 204: 392–400. Bibcode:2018RSEnv.204..392K. doi:10.1016/j.rse.2017.10.016. hdl: 2060/20180003069 . ISSN   0034-4257. PMC   7340154 . PMID   32636571.
  17. Hauhs, Michael; Lange, Holger (2008). "Classification of Runoff in Headwater Catchments: A Physical Problem?". Geography Compass. 2 (1): 235–254. doi:10.1111/j.1749-8198.2007.00075.x. ISSN   1749-8198.
  18. Liu, Meng; Liu, Dong; Liu, Le (2013-09-01). "Complexity research of regional groundwater depth series based on multiscale entropy: a case study of Jiangsanjiang Branch Bureau in China". Environmental Earth Sciences. 70 (1): 353–361. Bibcode:2013EES....70..353L. doi:10.1007/s12665-012-2132-y. ISSN   1866-6299. S2CID   128958458.
  19. Xing, Jing; Manning, Carol A. (April 2005). "Complexity and Automation Displays of Air Traffic Control: Literature Review and Analysis". Archived from the original on June 1, 2022.{{cite journal}}: Cite journal requires |journal= (help)
  20. Wang, Kang; Li, Li (November 2008). "Characterizing Heterogeneous Flow Patterns Using Information Measurements". 2008 First International Conference on Intelligent Networks and Intelligent Systems. pp. 654–657. doi:10.1109/ICINIS.2008.110. S2CID   8867649.
  21. Al Sawaf, Mohamad Basel; Kawanisi, Kiyosi (2020-11-01). "Assessment of mountain river streamflow patterns and flood events using information and complexity measures". Journal of Hydrology. 590: 125508. Bibcode:2020JHyd..59025508A. doi:10.1016/j.jhydrol.2020.125508. ISSN   0022-1694. S2CID   225261677.
  22. Javaheri Javid, Mohammad Ali; Alghamdi, Wajdi; Zimmer, Robert; al-Rifaie, Mohammad Majid (2016), Bi, Yaxin; Kapoor, Supriya; Bhatia, Rahul (eds.), "A Comparative Analysis of Detecting Symmetries in Toroidal Topology" (PDF), Intelligent Systems and Applications: Extended and Selected Results from the SAI Intelligent Systems Conference (IntelliSys) 2015, Studies in Computational Intelligence, Springer International Publishing, pp. 323–344, doi:10.1007/978-3-319-33386-1_16, ISBN   978-3-319-33386-1
  23. Jurado-González, Javier; Gómez-Barroso, José Luis (2022-11-28). "Economic complexity and Information Society paradigms: a hybrid contribution to explain economic growth". Technological and Economic Development of Economy. 28 (6): 1871–1896. doi: 10.3846/tede.2022.17104 . ISSN   2029-4921. S2CID   254339040.
  24. He, Kaijian; Lu, Xingjing; Zou, Yingchao; Keung Lai, Kin (2015-09-01). "Forecasting metal prices with a curvelet based multiscale methodology". Resources Policy. 45: 144–150. Bibcode:2015RePol..45..144H. doi:10.1016/j.resourpol.2015.03.011. ISSN   0301-4207.
  25. He, Kaijian; Xu, Yang; Zou, Yingchao; Tang, Ling (2015-05-01). "Electricity price forecasts using a Curvelet denoising based approach". Physica A: Statistical Mechanics and Its Applications. 425: 1–9. doi:10.1016/j.physa.2015.01.012. ISSN   0378-4371.
  26. Ahmed, Mosabber Uddin (2021), Ahad, Md Atiqur Rahman; Ahmed, Mosabber Uddin (eds.), "Complexity Analysis in Health Informatics", Signal Processing Techniques for Computational Health Informatics, Intelligent Systems Reference Library, Cham: Springer International Publishing, vol. 192, pp. 103–121, doi:10.1007/978-3-030-54932-9_4, ISBN   978-3-030-54932-9, S2CID   225129992 , retrieved 2021-02-01
  27. Shi Xiujian; Sun Zhiqiang; Li Long; Xie Hongwei (2009). "Human Cognitive Complexity Analysis in Transportation Systems". Logistics. Proceedings: 4361–4368. doi:10.1061/40996(330)637. ISBN   9780784409961.
  28. Zhang, Shutao; Qian, Jinwu; Shen, Linyong; Wu, Xi; Hu, Xiaowu (October 2015). "Gait complexity and frequency content analyses of patients with Parkinson's disease". 2015 International Symposium on Bioelectronics and Bioinformatics (ISBB). pp. 87–90. doi:10.1109/ISBB.2015.7344930. ISBN   978-1-4673-6609-0. S2CID   2891655.
  29. Wang, Jisung; Noh, Gyu-Jeong; Choi, Byung-Moon; Ku, Seung-Woo; Joo, Pangyu; Jung, Woo-Sung; Kim, Seunghwan; Lee, Heonsoo (2017-07-13). "Suppressed neural complexity during ketamine- and propofol-induced unconsciousness". Neuroscience Letters. 653: 320–325. doi:10.1016/j.neulet.2017.05.045. ISSN   0304-3940. PMID   28572032. S2CID   13767209.
  30. Bola, Michał; Orłowski, Paweł; Płomecka, Martyna; Marchewka, Artur (2019-01-30). "EEG signal diversity during propofol sedation: an increase in sedated but responsive, a decrease in sedated and unresponsive subjects". bioRxiv: 444281. doi:10.1101/444281. S2CID   214726084.
  31. Dilger, Alexander (2012-01-01). "Endogenous complexity, specialisation and general education". On the Horizon. 20 (1): 49–53. doi:10.1108/10748121211202062. ISSN   1074-8121.
  32. Ivanyuk, Vera Alekseevna (2015). "Dynamic strategic investment portfolio management model". elibrary.ru.
  33. Peña, Eric; Sayama, Hiroki (2021-05-02). "Life Worth Mentioning: Complexity in Life-Like Cellular Automata". Artificial Life . 27 (2): 105–112. doi: 10.1162/artl_a_00348 . PMID   34727158. S2CID   240424110.
  34. Javaheri Javid, Mohammad Ali (2019-11-30). Aesthetic Automata: Synthesis and Simulation of Aesthetic Behaviour in Cellular Automata (doctoral thesis). Goldsmiths, University of London. doi:10.25602/gold.00027681.