Local World Evolving Network Models

Last updated

Evolving networks are dynamic networks that change through time. In each period there are new nodes and edges that join the network while the old ones disappear. Such dynamic behaviour is characteristic for most real-world networks, regardless of their range - global or local. However, networks differ not only in their range but also in their topological structure. It is possible to distinguish:

Contents

One of the main feature which allows to differentiate networks is their evolution process. In random networks points are added and removed from the network in a totally random way (model of Erdős and Rényi). [1] Evolution of free scale networks is based on the preferential attachment – nodes connect to nodes that have already possessed a large number of links. In result hubs (nodes that have the largest number of edges) are created and networks follow power law of distribution (model of Barabási and Albert's [2] ). In opposite, in small world networks there are no hubs, and nodes are rather egalitarian and locally grouped in smaller clusters. These kind of networks are described by Watts and Strogatz (WS) model. [3] All aforementioned models assume that newly added points have a global information about the whole network. However, in case of large systems, such knowledge is rather rare. This strongly limits nodes’ possibilities of connection choice. As a result, decisions about links are made rather in a local world than in the whole network. Networks which consider this locality are called local-world networks and were first described by the Li and Chen model (2003). The local world model was extended inter alia by Gardeñes and Moreno (2004), Sen and Zhong, [4] Wen et al. [5] or Xuan et al. [6]

World Evolving Network Model of Li and Chen (2003)

The model starts with the set of small number of nodes and the small number of edges . There are M nodes that were selected randomly from the whole global network, so that they constitute a so-called “local world” for new coming nodes. Thus, every new node with m edges connects only to m existing nodes from its local world and does not link with nodes which are in the global system (the main difference from the BA model). In such case, the probability of connection may be defined as:

 

Where   and the term "Local-World" refers to all nodes, which are in interest of newly added node at time t. Thus, it may be rewritten:

 

while the dynamics are:

 

In every time t, it is true that  , so that two corner solutions are possible:   and  .

Fig.1.Degree distribution comparison in log-log scale of lower bound case with M = m = 3, and a local-world evolving network with M = 4 and m = 3. Networks have N = 10 000. The inset is in the log-linear scale of the same curves Degree distribution of the local-world evolving networks with M = 4; 10; 30, m = 3.jpg
Fig.1.Degree distribution comparison in log–log scale of lower bound case with M = m = 3, and a local-world evolving network with M = 4 and m = 3. Networks have N = 10 000. The inset is in the log-linear scale of the same curves

Case A. Lower bounded limit

A new node connects only to nodes from the initially chosen local world M. This identifies that in network growing process, preferential attachment (PA) selection is not efficient. The case is identical with BA scale free model, in which network grows without PA. The rate of change of the i th node’s degree may be written in the following way:

 

Thus, above proves that in the lower bound solution, network has an exponentially decayed degree distribution : (Fig.1)

Case B Lower bounded limit

In this case local world behaves in the same way as the global network. It evolves in time. Therefore, LW model may be compared to Barabasi–Albert scale-free model, and the rate of change of the 'i th' node’s degree may be expressed as:

 

This equality indicates that in the upper bound solution, LW model follows the degree distribution of the power law:  (Fig. 2)
Hence, from A and B, it may be found that among corner solutions, Li and Chen’s model represents a transition for the degree distribution between the exponential and the power-law (Fig.3).

New Local World Evolving Network Model of Sen and Zhong (2009)

The model is the extension of LM model in a sense that it divides nodes on these which have the information about the global network and on these which does not. To control for this diversification, parameter is introduced. Let be the ratio of the number of nodes obtaining the information about the global network to the total number of nodes. Because is a ratio, it must be that . When there is no nodes that ow the global information and NLW model comes down to the local-world network model. In turn, means that each node possesses the global information about the network, which makes NLW model identical with BA model.
The NWL model starts in the same way as LW – there is a set of small number of nodes m_0 and the small number of edges . There are M nodes that were selected randomly from the whole global network and established a “local world” for new coming nodes. However, in NLW model every new node with m edges can connect to global or local system. The decision depends on received information. If a new node gets information about the whole network, the probability that it will be connected with node i depends on the degree ki of that node, such that:

 

In turn, if the node was not provided in the global information and knows only its local world, it will link only with nodes from this system with the probability:

 

Thus, the general probability in the new local world model may be written as:

 

where is the probability that a new node possesses a knowledge about the global network. Similarly to the LW model, the NLW model distinguish three cases of local-world selection:

 ; and

The upper bound case (Case C) is the same as in the local world model.

Case A Lower bounded limit

In the lower limit there are only few nodes that meet holistic preferential attachment requirement, while most of them connect a new edge randomly. Moreover, the cumulative degree of the local world depends on the random selection. In such case, the dynamics of the system are described by:

 

with the assumption that: 
In this case, the degree distribution of the networks follows a power-low distribution, and the exponent of the scale-free network equals so that the initial assumption about small indicates that the power-low exponent of the network reaches a high value.

Case B.

In time t there are nodes If the new coming node does not have the information about the global network, it will link to i node in the local system with the probability . Thus, the dynamics may be written as follows:

 

with the assumption that: 
As in previous case, the evolving network has a power-law degree distribution, however, with larger γ exponent, which equals : 
It may be noticed that the ratio is the only parameter of the scale-free exponent of the new model. Thus, the significant improvement of the model comes from the introduction of , which by adding or removing nodes that possess the information about the global network, allows to control a topological structure of a network.

Related Research Articles

<span class="mw-page-title-main">Fokker–Planck equation</span> Partial differential equation

In statistical mechanics and information theory, the Fokker–Planck equation is a partial differential equation that describes the time evolution of the probability density function of the velocity of a particle under the influence of drag forces and random forces, as in Brownian motion. The equation can be generalized to other observables as well. The Fokker-Planck equation has multiple applications in information theory, graph theory, data science, finance, economics etc.

<span class="mw-page-title-main">Path integral formulation</span> Formulation of quantum mechanics

The path integral formulation is a description in quantum mechanics that generalizes the stationary action principle of classical mechanics. It replaces the classical notion of a single, unique classical trajectory for a system with a sum, or functional integral, over an infinity of quantum-mechanically possible trajectories to compute a quantum amplitude.

In information geometry, the Fisher information metric is a particular Riemannian metric which can be defined on a smooth statistical manifold, i.e., a smooth manifold whose points are probability measures defined on a common probability space. It can be used to calculate the informational difference between measurements.

<span class="mw-page-title-main">Boltzmann machine</span> Type of stochastic recurrent neural network

A Boltzmann machine is a stochastic spin-glass model with an external field, i.e., a Sherrington–Kirkpatrick model, that is a stochastic Ising model. It is a statistical physics technique applied in the context of cognitive science. It is also classified as a Markov random field.

<span class="mw-page-title-main">Markov random field</span> Set of random variables

In the domain of physics and probability, a Markov random field (MRF), Markov network or undirected graphical model is a set of random variables having a Markov property described by an undirected graph. In other words, a random field is said to be a Markov random field if it satisfies Markov properties. The concept originates from the Sherrington–Kirkpatrick model.

<span class="mw-page-title-main">Backpropagation</span> Optimization algorithm for artificial neural networks

As a machine-learning algorithm, backpropagation performs a backward pass to adjust a neural network model's parameters, aiming to minimize error. In a multi-layered network, backpropagation uses the following steps:

  1. Propagate training data through the model from input to predicted output by computing the successive hidden layers' outputs and finally the final layer's output.
  2. Adjust the model weights to reduce the error relative to the weights.
    1. The error is typically the squared difference between prediction and target.
    2. For each weight, the slope or derivative of the error is found, and the weight adjusted by a negative multiple of this derivative, so as to go downslope toward the minimum-error configuration.
    3. This derivative is easy to calculate for final layer weights, and possible to calculate for one layer given the next layer's derivatives. Starting at the end, then, the derivatives are calculated layer by layer toward the beginning -- thus "backpropagation".
  3. Repeatedly update the weights until they converge or the model has undergone enough iterations.
<span class="mw-page-title-main">Degree distribution</span>

In the study of graphs and networks, the degree of a node in a network is the number of connections it has to other nodes and the degree distribution is the probability distribution of these degrees over the whole network.

The birth–death process is a special case of continuous-time Markov process where the state transitions are of only two types: "births", which increase the state variable by one and "deaths", which decrease the state by one. It was introduced by William Feller. The model's name comes from a common application, the use of such models to represent the current size of a population where the transitions are literal births and deaths. Birth–death processes have many applications in demography, queueing theory, performance engineering, epidemiology, biology and other areas. They may be used, for example, to study the evolution of bacteria, the number of people with a disease within a population, or the number of customers in line at the supermarket.

In queueing theory, a discipline within the mathematical theory of probability, a Jackson network is a class of queueing network where the equilibrium distribution is particularly simple to compute as the network has a product-form solution. It was the first significant development in the theory of networks of queues, and generalising and applying the ideas of the theorem to search for similar product-form solutions in other networks has been the subject of much research, including ideas used in the development of the Internet. The networks were first identified by James R. Jackson and his paper was re-printed in the journal Management Science’s ‘Ten Most Influential Titles of Management Sciences First Fifty Years.’

<span class="mw-page-title-main">Watts–Strogatz model</span> Method of generating random small-world graphs

The Watts–Strogatz model is a random graph generation model that produces graphs with small-world properties, including short average path lengths and high clustering. It was proposed by Duncan J. Watts and Steven Strogatz in their article published in 1998 in the Nature scientific journal. The model also became known as the (Watts) beta model after Watts used to formulate it in his popular science book Six Degrees.

<span class="mw-page-title-main">Gaussian network model</span>

The Gaussian network model (GNM) is a representation of a biological macromolecule as an elastic mass-and-spring network to study, understand, and characterize the mechanical aspects of its long-time large-scale dynamics. The model has a wide range of applications from small proteins such as enzymes composed of a single domain, to large macromolecular assemblies such as a ribosome or a viral capsid. Protein domain dynamics plays key roles in a multitude of molecular recognition and cell signalling processes. Protein domains, connected by intrinsically disordered flexible linker domains, induce long-range allostery via protein domain dynamics. The resultant dynamic modes cannot be generally predicted from static structures of either the entire protein or individual domains.

<span class="mw-page-title-main">Network science</span> Academic field

Network science is an academic field which studies complex networks such as telecommunication networks, computer networks, biological networks, cognitive and semantic networks, and social networks, considering distinct elements or actors represented by nodes and the connections between the elements or actors as links. The field draws on theories and methods including graph theory from mathematics, statistical mechanics from physics, data mining and information visualization from computer science, inferential modeling from statistics, and social structure from sociology. The United States National Research Council defines network science as "the study of network representations of physical, biological, and social phenomena leading to predictive models of these phenomena."

In fracture mechanics, the energy release rate, , is the rate at which energy is transformed as a material undergoes fracture. Mathematically, the energy release rate is expressed as the decrease in total potential energy per increase in fracture surface area, and is thus expressed in terms of energy per unit area. Various energy balances can be constructed relating the energy released during fracture to the energy of the resulting new surface, as well as other dissipative processes such as plasticity and heat generation. The energy release rate is central to the field of fracture mechanics when solving problems and estimating material properties related to fracture and fatigue.

<span class="mw-page-title-main">Diffusion</span> Transport of dissolved species from the highest to the lowest concentration region

Diffusion is the net movement of anything generally from a region of higher concentration to a region of lower concentration. Diffusion is driven by a gradient in Gibbs free energy or chemical potential. It is possible to diffuse "uphill" from a region of lower concentration to a region of higher concentration, as in spinodal decomposition. Diffusion is a stochastic process due to the inherent randomness of the diffusing entity and can be used to model many real-life stochastic scenarios. Therefore, diffusion and the corresponding mathematical models are used in several fields beyond physics, such as statistics, probability theory, information theory, neural networks, finance, and marketing.

Simultaneous perturbation stochastic approximation (SPSA) is an algorithmic method for optimizing systems with multiple unknown parameters. It is a type of stochastic approximation algorithm. As an optimization method, it is appropriately suited to large-scale population models, adaptive modeling, simulation optimization, and atmospheric modeling. Many examples are presented at the SPSA website http://www.jhuapl.edu/SPSA. A comprehensive book on the subject is Bhatnagar et al. (2013). An early paper on the subject is Spall (1987) and the foundational paper providing the key theory and justification is Spall (1992).

<span class="mw-page-title-main">Louvain method</span> Clustering and community detection algorithm

The Louvain method for community detection is a method to extract non-overlapping communities from large networks created by Blondel et al. from the University of Louvain. The method is a greedy optimization method that appears to run in time where is the number of nodes in the network.

A two-dimensional conformal field theory is a quantum field theory on a Euclidean two-dimensional space, that is invariant under local conformal transformations.

<span class="mw-page-title-main">Configuration model</span>

In network science, the configuration model is a method for generating random networks from a given degree sequence. It is widely used as a reference model for real-life social networks, because it allows the modeler to incorporate arbitrary degree distributions.

<span class="mw-page-title-main">Hyperbolastic functions</span> Mathematical functions

The hyperbolastic functions, also known as hyperbolastic growth models, are mathematical functions that are used in medical statistical modeling. These models were originally developed to capture the growth dynamics of multicellular tumor spheres, and were introduced in 2005 by Mohammad Tabatabai, David Williams, and Zoran Bursac. The precision of hyperbolastic functions in modeling real world problems is somewhat due to their flexibility in their point of inflection. These functions can be used in a wide variety of modeling problems such as tumor growth, stem cell proliferation, pharma kinetics, cancer growth, sigmoid activation function in neural networks, and epidemiological disease progression or regression.

In computational and mathematical biology, a biological lattice-gas cellular automaton (BIO-LGCA) is a discrete model for moving and interacting biological agents, a type of cellular automaton. The BIO-LGCA is based on the lattice-gas cellular automaton (LGCA) model used in fluid dynamics. A BIO-LGCA model describes cells and other motile biological agents as point particles moving on a discrete lattice, thereby interacting with nearby particles. Contrary to classic cellular automaton models, particles in BIO-LGCA are defined by their position and velocity. This allows to model and analyze active fluids and collective migration mediated primarily through changes in momentum, rather than density. BIO-LGCA applications include cancer invasion and cancer progression.

References

  1. Erdős,P. and A.Rényi (1961). On the evolution of random graphs, Publ. Math. Inst. Hung. Acad. Sci,Vol.5, p.17-61
  2. Albert, R. and A.L.Barabasi (2000). Physical Review Letters, Vol.85, No.24, p.5234
  3. Watts,J.D. and H.S.Strogatz (1998). Collective dynamics of 'small-world' networks, Nature, Vol.393, p.440-442
  4. Sen,Q. and D.G.Zhong ()Chinese Physics B, Vol.18, No.2, p.383
  5. Wen,G., Z.Duan, G.Chen G and X.Geng (2011). Physica A, Vol.390, p.4012
  6. Xuan,Q., Y.Li and T.Wu (2007). Physica A, Vol.378, p.561
  7. Li,X. and G.R.Chen (2003). Physica A, Vol. 328, p. 274

Sources