Network entropy

Last updated

In network science, the network entropy is a disorder measure derived from information theory to describe the level of randomness and the amount of information encoded in a graph. [1] It is a relevant metric to quantitatively characterize real complex networks and can also be used to quantify network complexity [1] [2]

Contents

Formulations

According to a publication from Entropy, a journal published by MDPI, there are several formulations in which to measure the network entropy and, as a rule, they all require a particular property of the graph to be focused, such as the adjacency matrix, degree sequence, degree distribution or number of bifurcations, what might lead to values of entropy that aren't invariant to the chosen network description. [3]

Degree Distribution Shannon Entropy

The Shannon entropy can be measured for the network degree probability distribution as an average measurement of the heterogeneity of the network.

This formulation has limited use with regards to complexity, information content, causation and temporal information. Be that as it may, algorithmic complexity has the ability to characterize any general or universal property of a graph or network and it is proven that graphs with low entropy have low algorithmic complexity because the statistical regularities found in a graph are useful for computer programs to recreate it. The same cannot be said for high entropy networks though, as these might have any value for algorithmic complexity. [3]

Random Walker Shannon Entropy

Due to the limits of the previous formulation, it is possible to take a different approach while keeping the usage of the original Shannon Entropy equation.

Consider a random walker that travels around the graph, going from a node to any node adjacent to with equal probability. The probability distribution that describes the behavior of this random walker would thus be

,

where is the graph adjacency matrix and is the node degree.

From that, the Shannon entropy from each node can be defined as

and, since , the normalized node entropy is calculated

This leads to a normalized network entropy , calculated by averaging the normalized node entropy over the whole network [4]

The normalized network entropy is maximal when the network is fully connected and decreases the sparser the network becomes . Notice that isolated nodes do not have its probability defined and, therefore, are not considered when measuring the network entropy. This formulation of network entropy has low sensitivity to hubs due to the logarithmic factor and is more meaningful for weighted networks., [4] what ultimately makes it hard to differentiate scale-free networks using this measure alone. [2]

Random Walker Kolmogorov–Sinai Entropy

The limitations of the random walker Shannon entropy can be overcome by adapting it to use a Kolmogorov–Sinai entropy. In this context, network entropy is the entropy of a stochastic matrix associated with the graph adjacency matrix and the random walker Shannon entropy is called the dynamic entropy of the network. From that, let be the dominant eigenvalue of . It is proven that satisfies a variational principal [5] that is equivalent to the dynamic entropy for unweighted networks, i.e., the adjacency matrix consists exclusively of boolean values. Therefore, the topological entropy is defined as

This formulation is important to the study of network robustness, i.e., the capacity of the network to withstand random structural changes. Robustness is actually difficult to be measured numerically whereas the entropy can be easily calculated for any network, which is especially important in the context of non-stationary networks. The entropic fluctuation theorem shows that this entropy is positively correlated to robustness and hence a greater insensitivity of an observable to dynamic or structural perturbations of the network. Moreover, the eigenvalues are inherently related to the multiplicity of internal pathways, leading to a negative correlation between the topological entropy and the shortest average path length. [6]

Other than that, the Kolmogorov entropy is related to the Ricci curvature of the network, [7] a metric that has been used to differentiate stages of cancer from gene co-expression networks, [8] as well as to give hallmarks of financial crashes from stock correlation networks [9]

Von Neumann entropy

Von Neumann entropy is the extension of the classical Gibbs entropy in a quantum context. This entropy is constructed from a density matrix : historically, the first proposed candidate for such a density matrix has been an expression of the Laplacian matrix L associated with the network. The average von Neumann entropy of an ensemble is calculated as: [10]

For random network ensemble , the relation between and is nonmonotonic when the average connectivity is varied.

For canonical power-law network ensembles, the two entropies are linearly related. [11]

Networks with given expected degree sequences suggest that, heterogeneity in the expected degree distribution implies an equivalence between a quantum and a classical description of networks, which respectively corresponds to the von Neumann and the Shannon entropy. [12]

This definition of the Von Neumann entropy can also be extended to multilayer networks with tensorial approach [13] and has been used successfully to reduce their dimensionality from a structural point of perspective. [14]

However, it has been shown that this definition of entropy does not satisfy the property of sub-additivity (see Von Neumann entropy's subadditivity), expected to hold theoretically. A more grounded definition, satisfying this fundamental property, has been introduced by Manlio De Domenico and Biamonte [15] as a quantum-like Gibbs state

where

is a normalizing factor which plays the role of the partition function, and is a tunable parameter which allows multi-resolution analysis. If is interpreted as a temporal parameter, this density matrix is formally proportional to the propagator of a diffusive process on the top of the network.

This feature has been used to build a statistical field theory of complex information dynamics, where the density matrix can be interpreted in terms of the super-position of streams operators whose action is to activate information flows among nodes. [16] The framework has been successfully applied to analyze the protein-protein interaction networks of virus-human interactomes, including the SARS-CoV-2, to unravel the systemic features of infection of the latter at microscopic, mesoscopic and macroscopic scales, [17] as well as to assess the importance of nodes for integrating information flows within the network and the role they play in network robustness. [18]

This approach has been generalized to deal with other types of dynamics, such as random walks, on the top of multilayer networks, providing an effective way to reduce the dimensionality of such systems without altering their structure. [19] Using both classical and maximum-entropy random walks, the corresponding density matrices have been used to encode the network states of the human brain and to assess, at multiple scales, connectome’s information capacity at different stages of dementia. [20]

Maximum Entropy Principle

The maximum entropy principle is a variational principal stating that the probability distribution best representing the current state of a system is the one which maximizes the Shannon entropy. [21] This concept can be used to generate an ensemble of random graphs with given structural properties derived from the maximum entropy approach which, in its turn, describes the most probable network configuration: the maximum entropy principle allows for maximally unbiased information when lacking complete knowledge (microscopic configuration is not accessible, e.g.: we don't know the adjacency matrix). On the other hand, this ensemble serves as a null model when the actual microscopic configuration of the network is known, allowing to assess the significance of empirical patterns found in the network [22]

Complex Network Ensembles

It is possible to extend the network entropy formulations to instead measure the ensemble entropy. Let be an ensemble of all graphs with the same number of nodes and type of links, is defined as the occurrence probability of a graph . From the maximum entropy principle, it is desired to achieve such that it maximizes the Shannon entropy

Moreover, constraints shall be imposed in order to extract conclusions. Soft constraints (constraints set over the whole ensemble ) will lead to the canonical ensemble whereas hard constraints (constraints set over each graph ) are going to define the microcanonical ensemble. In the end, these ensembles lead to models that can validate a range of network patterns and even reconstruct graphs from limited knowledge, showing that maximum entropy models based on local constraints can quantify the relevance of a set of observed features and extracting meaningful information from huge big data streams in real-time and high-dimensional noisy data. [22]

See also Entropy of network ensembles.

Related Research Articles

<span class="mw-page-title-main">Random graph</span> Graph generated by a random process

In mathematics, random graph is the general term to refer to probability distributions over graphs. Random graphs may be described simply by a probability distribution, or by a random process which generates them. The theory of random graphs lies at the intersection between graph theory and probability theory. From a mathematical perspective, random graphs are used to answer questions about the properties of typical graphs. Its practical applications are found in all areas in which complex networks need to be modeled – many random graph models are thus known, mirroring the diverse types of complex networks encountered in different areas. In a mathematical context, random graph refers almost exclusively to the Erdős–Rényi random graph model. In other contexts, any graph model may be referred to as a random graph.

In graph theory, a clustering coefficient is a measure of the degree to which nodes in a graph tend to cluster together. Evidence suggests that in most real-world networks, and in particular social networks, nodes tend to create tightly knit groups characterised by a relatively high density of ties; this likelihood tends to be greater than the average probability of a tie randomly established between two nodes.

<span class="mw-page-title-main">Centrality</span> Degree of connectedness within a graph

In graph theory and network analysis, indicators of centrality assign numbers or rankings to nodes within a graph corresponding to their network position. Applications include identifying the most influential person(s) in a social network, key infrastructure nodes in the Internet or urban networks, super-spreaders of disease, and brain networks. Centrality concepts were first developed in social network analysis, and many of the terms used to measure centrality reflect their sociological origin.

In probability theory and mathematical physics, a random matrix is a matrix-valued random variable—that is, a matrix in which some or all elements are random variables. Many important properties of physical systems can be represented mathematically as matrix problems. For example, the thermal conductivity of a lattice can be computed from the dynamical matrix of the particle-particle interactions within the lattice.

<span class="mw-page-title-main">Degree distribution</span>

In the study of graphs and networks, the degree of a node in a network is the number of connections it has to other nodes and the degree distribution is the probability distribution of these degrees over the whole network.

<span class="mw-page-title-main">Watts–Strogatz model</span> Method of generating random small-world graphs

The Watts–Strogatz model is a random graph generation model that produces graphs with small-world properties, including short average path lengths and high clustering. It was proposed by Duncan J. Watts and Steven Strogatz in their article published in 1998 in the Nature scientific journal. The model also became known as the (Watts) beta model after Watts used to formulate it in his popular science book Six Degrees.

Mixing patterns refer to systematic tendencies of one type of nodes in a network to connect to another type. For instance, nodes might tend to link to others that are very similar or very different. This feature is common in many social networks, although it also appears sometimes in non-social networks. Mixing patterns are closely related to assortativity; however, for the purposes of this article, the term is used to refer to assortative or disassortative mixing based on real-world factors, either topological or sociological.

<span class="mw-page-title-main">Random geometric graph</span> In graph theory, the mathematically simplest spatial network

In graph theory, a random geometric graph (RGG) is the mathematically simplest spatial network, namely an undirected graph constructed by randomly placing N nodes in some metric space and connecting two nodes by a link if and only if their distance is in a given range, e.g. smaller than a certain neighborhood radius, r.

<span class="mw-page-title-main">Network science</span> Academic field

Network science is an academic field which studies complex networks such as telecommunication networks, computer networks, biological networks, cognitive and semantic networks, and social networks, considering distinct elements or actors represented by nodes and the connections between the elements or actors as links. The field draws on theories and methods including graph theory from mathematics, statistical mechanics from physics, data mining and information visualization from computer science, inferential modeling from statistics, and social structure from sociology. The United States National Research Council defines network science as "the study of network representations of physical, biological, and social phenomena leading to predictive models of these phenomena."

Algebraic signal processing (ASP) is an emerging area of theoretical signal processing (SP). In the algebraic theory of signal processing, a set of filters is treated as an (abstract) algebra, a set of signals is treated as a module or vector space, and convolution is treated as an algebra representation. The advantage of algebraic signal processing is its generality and portability.

<span class="mw-page-title-main">Google matrix</span> Stochastic matrix representing links between entities

A Google matrix is a particular stochastic matrix that is used by Google's PageRank algorithm. The matrix represents a graph with edges representing links between pages. The PageRank of each page can then be generated iteratively from the Google matrix using the power method. However, in order for the power method to converge, the matrix must be stochastic, irreducible and aperiodic.

<span class="mw-page-title-main">Exponential family random graph models</span>

Exponential Random Graph Models (ERGMs) are a family of statistical models for analyzing data from social and other networks. Examples of networks examined using ERGM include knowledge networks, organizational networks, colleague networks, social media networks, networks of scientific development, and others.

<span class="mw-page-title-main">Hyperbolic geometric graph</span>

A hyperbolic geometric graph (HGG) or hyperbolic geometric network (HGN) is a special type of spatial network where (1) latent coordinates of nodes are sprinkled according to a probability density function into a hyperbolic space of constant negative curvature and (2) an edge between two nodes is present if they are close according to a function of the metric (typically either a Heaviside step function resulting in deterministic connections between vertices closer than a certain threshold distance, or a decaying function of hyperbolic distance yielding the connection probability). A HGG generalizes a random geometric graph (RGG) whose embedding space is Euclidean.

<span class="mw-page-title-main">Multidimensional network</span> Networks with multiple kinds of relations

In network theory, multidimensional networks, a special type of multilayer network, are networks with multiple kinds of relations. Increasingly sophisticated attempts to model real-world systems as multidimensional networks have yielded valuable insight in the fields of social network analysis, economics, urban and international transport, ecology, psychology, medicine, biology, commerce, climatology, physics, computational neuroscience, operations management, and finance.

<span class="mw-page-title-main">Efficiency (network science)</span>

In network science, the efficiency of a network is a measure of how efficiently it exchanges information and it is also called communication efficiency. The underlying idea is that the more distant two nodes are in the network, the less efficient their communication will be. The concept of efficiency can be applied to both local and global scales in a network. On a global scale, efficiency quantifies the exchange of information across the whole network where information is concurrently exchanged. The local efficiency quantifies a network's resistance to failure on a small scale. That is the local efficiency of a node characterizes how well information is exchanged by its neighbors when it is removed.

<span class="mw-page-title-main">Biased random walk on a graph</span> Structural analysis of a network

In network science, a biased random walk on a graph is a time path process in which an evolving variable jumps from its current state to one of various potential new states; unlike in a pure random walk, the probabilities of the potential new states are unequal.

Maximal entropy random walk (MERW) is a popular type of biased random walk on a graph, in which transition probabilities are chosen accordingly to the principle of maximum entropy, which says that the probability distribution which best represents the current state of knowledge is the one with largest entropy. While standard random walk chooses for every vertex uniform probability distribution among its outgoing edges, locally maximizing entropy rate, MERW maximizes it globally by assuming uniform probability distribution among all paths in a given graph.

<span class="mw-page-title-main">Configuration model</span>

In network science, the configuration model is a method for generating random networks from a given degree sequence. It is widely used as a reference model for real-life social networks, because it allows the modeler to incorporate arbitrary degree distributions.

<span class="mw-page-title-main">Maximum-entropy random graph model</span>

Maximum-entropy random graph models are random graph models used to study complex networks subject to the principle of maximum entropy under a set of structural constraints, which may be global, distributional, or local.

A set of networks that satisfies given structural characteristics can be treated as a network ensemble. Brought up by Ginestra Bianconi in 2007, the entropy of a network ensemble measures the level of the order or uncertainty of a network ensemble.

References

  1. 1 2 Anand, Kartik; Krioukov, Dmitri; Bianconi, Ginestra (2014). "Entropy distribution and condensation in random networks with a given degree distribution". Physical Review E. 89 (6): 062807. arXiv: 1403.5884 . Bibcode:2014PhRvE..89f2807A. doi:10.1103/PhysRevE.89.062807. PMID   25019833. S2CID   761765.
  2. 1 2 Freitas, Cristopher GS; Aquino, Andre LL; Ramos, Heitor S; Frery, Alejandro C; Rosso, Osvaldo A (2019). "A detailed characterization of complex networks using Information Theory". Scientific Reports. 9 (1): 16689. Bibcode:2019NatSR...916689F. doi:10.1038/s41598-019-53167-5. PMC   6853913 . PMID   31723172. S2CID   207987035.
  3. 1 2 Zenil, Hector; Kiani, Narsis A; Tegnér, Jesper (2018). "A review of graph and network complexity from an algorithmic information perspective". Entropy. 20 (8): 551. Bibcode:2018Entrp..20..551Z. doi: 10.3390/e20080551 . PMC   7513075 . PMID   33265640.
  4. 1 2 Small, Michael (2013). "Complex networks from time series: Capturing dynamics". 2013 IEEE International Symposium on Circuits and Systems (ISCAS2013). pp. 2509–2512. doi:10.1109/ISCAS.2013.6572389. ISBN   978-1-4673-5762-3. S2CID   9275909.
  5. Arnold, Ludwig; Gundlach, Volker Matthias; Demetrius, Lloyd (1994). "Evolutionary formalism for products of positive random matrices". The Annals of Applied Probability. 4 (3): 859–901. doi: 10.1214/aoap/1177004975 . JSTOR   2245067.
  6. Demetrius, Lloyd; Manke, Thomas (2005). "Robustness and network evolution—an entropic principle". Physica A: Statistical Mechanics and Its Applications. 346 (3): 682–696. Bibcode:2005PhyA..346..682D. doi:10.1016/j.physa.2004.07.011.
  7. Lott, J.; Villani, C. (2009). "Ricci curvature for metric-measure spaces via optimal transport". Annals of Mathematics. 169 (3): 903–991. arXiv: math/0412127 . doi:10.4007/annals.2009.169.903. S2CID   15556613.
  8. Sandhu, R.; Georgiou, T.; Reznik, E.; Zhu, L.; Kolesov, I.; Senbabaoglu, Y.; Tannenbaum, A. (2015). "Graph curvature for differentiating cancer networks". Scientific Reports. 5: 12323. Bibcode:2015NatSR...512323S. doi:10.1038/srep12323. PMC   4500997 . PMID   26169480.
  9. Sandhu, Romeil S; Georgiou, Tryphon T; Tannenbaum, Allen R (2016). "Ricci curvature: An economic indicator for market fragility and systemic risk". Science Advances. 2 (5): e1501495. Bibcode:2016SciA....2E1495S. doi:10.1126/sciadv.1501495. PMC   4928924 . PMID   27386522.
  10. Du, Wenxue; Li, Xueliang; Li, Yiyang; Severini, Simone (30 December 2010). "A note on the von Neumann entropy of random graphs". Linear Algebra and Its Applications. 433 (11): 1722–1725. doi: 10.1016/j.laa.2010.06.040 . ISSN   0024-3795.
  11. Anand, Kartik; Bianconi, Ginestra (13 October 2009). "Entropy measures for networks: Toward an information theory of complex topologies". Physical Review E. 80 (4): 045102. arXiv: 0907.1514 . Bibcode:2009PhRvE..80d5102A. doi:10.1103/PhysRevE.80.045102. PMID   19905379. S2CID   27419558.
  12. Anand, Kartik; Bianconi, Ginestra; Severini, Simone (18 March 2011). "Shannon and von Neumann entropy of random networks with heterogeneous expected degree". Physical Review E. 83 (3): 036109. arXiv: 1011.1565 . Bibcode:2011PhRvE..83c6109A. doi:10.1103/PhysRevE.83.036109. PMID   21517560. S2CID   1482301.
  13. De Domenico, Manlio; Solé-Ribalta, Albert; Cozzo, Emanuele; Kivelä, Mikko; Moreno, Yamir; Porter, Mason A.; Gómez, Sergio; Arenas, Alex (4 December 2013). "Mathematical Formulation of Multilayer Networks". Physical Review X. 3 (4): 041022. arXiv: 1307.4977 . Bibcode:2013PhRvX...3d1022D. doi:10.1103/PhysRevX.3.041022. S2CID   16611157.
  14. De Domenico, Manlio; Nicosia, Vincenzo; Arenas, Alex; Latora, Vito (23 April 2015). "Structural reducibility of multilayer networks" (PDF). Nature Communications. 6: 6864. Bibcode:2015NatCo...6.6864D. doi:10.1038/ncomms7864. PMID   25904309. S2CID   16776349.
  15. De Domenico, Manlio; Biamonte, Jacob (21 December 2016). "Spectral Entropies as Information-Theoretic Tools for Complex Network Comparison". Physical Review X. 6 (4): 041062. arXiv: 1609.01214 . Bibcode:2016PhRvX...6d1062D. doi:10.1103/PhysRevX.6.041062. S2CID   51786781.
  16. Ghavasieh, Arsham; Nicolini, Carlo; De Domenico, Manlio (10 November 2020). "Statistical physics of complex information dynamics". Physical Review E. 102 (5): 052304. arXiv: 2010.04014 . Bibcode:2020PhRvE.102e2304G. doi:10.1103/PhysRevE.102.052304. PMID   33327131. S2CID   222208856.
  17. Ghavasieh, Arsham; Bontorin, Sebastiano; Artime, Oriol; Verstraete, Nina; De Domenico, Manlio (23 April 2021). "Multiscale statistical physics of the pan-viral interactome unravels the systemic nature of SARS-CoV-2 infections". Communications Physics. 4 (1): 83. Bibcode:2021CmPhy...4...83G. doi: 10.1038/s42005-021-00582-8 .
  18. Ghavasieh, Arsham; Stella, Massimo; Biamonte, Jacob; De Domenico, Manlio (10 June 2021). "Unraveling the effects of multiscale network entanglement on empirical systems". Communications Physics. 4 (1): 129. arXiv: 2008.05368 . Bibcode:2021CmPhy...4..129G. doi:10.1038/s42005-021-00633-0. S2CID   221104066.
  19. Ghavasieh, Arsham; De Domenico, Manlio (13 February 2020). "Enhancing transport properties in interconnected systems without altering their structure". Physical Review Research. 2 (1): 13–15. arXiv: 2001.04450 . Bibcode:2020PhRvR...2a3155G. doi:10.1103/PhysRevResearch.2.013155. S2CID   210165034.
  20. Benigni, Barbara; Ghavasieh, Arsham; Corso, Alessandra; D'Andrea, Valeria; De Domenico, Manlio (22 June 2021). "Persistence of information flow: a multiscale characterization of human brain". Network Neuroscience. 5 (3): 831–850. doi: 10.1162/netn_a_00203 . PMC   8567833 . PMID   34746629.
  21. Jaynes, E. T. (1957). "Information Theory and Statistical Mechanics" (PDF). Physical Review. Series II. 106 (4): 620–630. Bibcode:1957PhRv..106..620J. doi:10.1103/PhysRev.106.620. MR   0087305. S2CID   17870175.
  22. 1 2 Cimini, Giulio; Squartini, Tiziano; Saracco, Fabio; Garlaschelli, Diego; Gabrielli, Andrea; Caldarelli, Guido (2019). "The statistical physics of real-world networks". Nature Reviews Physics. 1 (1): 58–71. arXiv: 1810.05095 . Bibcode:2019NatRP...1...58C. doi:10.1038/s42254-018-0002-6. S2CID   52963395.