Eigenvector centrality

Last updated

In graph theory, eigenvector centrality (also called eigencentrality or prestige score [1] ) is a measure of the influence of a node in a connected network. Relative scores are assigned to all nodes in the network based on the concept that connections to high-scoring nodes contribute more to the score of the node in question than equal connections to low-scoring nodes. A high eigenvector score means that a node is connected to many nodes who themselves have high scores. [2] [3]

Contents

Google's PageRank and the Katz centrality are variants of the eigenvector centrality. [4]

Using the adjacency matrix to find eigenvector centrality

For a given graph with vertices let be the adjacency matrix, i.e. if vertex is linked to vertex , and otherwise. The relative centrality score, , of vertex can be defined as:

where is the set of neighbors of and is a constant. With a small rearrangement this can be rewritten in vector notation as the eigenvector equation

In general, there will be many different eigenvalues for which a non-zero eigenvector solution exists. However, the connectedness assumption and the additional requirement that all the entries in the eigenvector be non-negative imply (by the Perron–Frobenius theorem) that only the greatest eigenvalue results in the desired centrality measure. [5] The component of the related eigenvector then gives the relative centrality score of the vertex in the network. The eigenvector is only defined up to a common factor, so only the ratios of the centralities of the vertices are well defined. To define an absolute score, one must normalise the eigenvector e.g. such that the sum over all vertices is 1 or the total number of vertices n. Power iteration is one of many eigenvalue algorithms that may be used to find this dominant eigenvector. [4] Furthermore, this can be generalized so that the entries in A can be real numbers representing connection strengths, as in a stochastic matrix.

Normalized eigenvector centrality scoring

Google's PageRank is based on the normalized eigenvector centrality, or normalized prestige, combined with a random jump assumption. [1] The PageRank of a node has recursive dependence on the PageRank of other nodes that point to it. The normalized adjacency matrix is defined as:

where is the out-degree of node , or in vector form:

,

where is the vector of ones, and is the diagonal matrix of vector . is a row-stochastic matrix.

The normalized eigenvector prestige score is defined as:

or in vector form,

Applications

Eigenvector centrality is a measure of the influence a node has on a network. If a node is pointed to by many nodes (which also have high eigenvector centrality) then that node will have high eigenvector centrality. [6]

The earliest use of eigenvector centrality is by Edmund Landau in an 1895 paper on scoring chess tournaments. [7] [8]

More recently, researchers across many fields have analyzed applications, manifestations, and extensions of eigenvector centrality in a variety of domains:

See also

Related Research Articles

In graph theory, an expander graph is a sparse graph that has strong connectivity properties, quantified using vertex, edge or spectral expansion. Expander constructions have spawned research in pure and applied mathematics, with several applications to complexity theory, design of robust computer networks, and the theory of error-correcting codes.

<span class="mw-page-title-main">Principal component analysis</span> Method of data analysis

Principal component analysis (PCA) is a linear dimensionality reduction technique with applications in exploratory data analysis, visualization and data preprocessing.

In linear algebra, a diagonal matrix is a matrix in which the entries outside the main diagonal are all zero; the term usually refers to square matrices. Elements of the main diagonal can either be zero or nonzero. An example of a 2×2 diagonal matrix is , while an example of a 3×3 diagonal matrix is. An identity matrix of any size, or any multiple of it is a diagonal matrix called scalar matrix, for example, . In geometry, a diagonal matrix may be used as a scaling matrix, since matrix multiplication with it results in changing scale (size) and possibly also shape; only a scalar matrix results in uniform change in scale.

In mathematics, a Hermitian matrix is a complex square matrix that is equal to its own conjugate transpose—that is, the element in the i-th row and j-th column is equal to the complex conjugate of the element in the j-th row and i-th column, for all indices i and j:

In graph theory and computer science, an adjacency matrix is a square matrix used to represent a finite graph. The elements of the matrix indicate whether pairs of vertices are adjacent or not in the graph.

<span class="mw-page-title-main">Eigenface</span> Set of eigenvectors used in the computer vision problem of human face recognition

An eigenface is the name given to a set of eigenvectors when used in the computer vision problem of human face recognition. The approach of using eigenfaces for recognition was developed by Sirovich and Kirby and used by Matthew Turk and Alex Pentland in face classification. The eigenvectors are derived from the covariance matrix of the probability distribution over the high-dimensional vector space of face images. The eigenfaces themselves form a basis set of all images used to construct the covariance matrix. This produces dimension reduction by allowing the smaller set of basis images to represent the original training images. Classification can be achieved by comparing how faces are represented by the basis set.

In numerical analysis, one of the most important problems is designing efficient and stable algorithms for finding the eigenvalues of a matrix. These eigenvalue algorithms may also find eigenvectors.

In mathematics, spectral graph theory is the study of the properties of a graph in relationship to the characteristic polynomial, eigenvalues, and eigenvectors of matrices associated with the graph, such as its adjacency matrix or Laplacian matrix.

<span class="mw-page-title-main">Euler's rotation theorem</span> Movement with a fixed point is rotation

In geometry, Euler's rotation theorem states that, in three-dimensional space, any displacement of a rigid body such that a point on the rigid body remains fixed, is equivalent to a single rotation about some axis that runs through the fixed point. It also means that the composition of two rotations is also a rotation. Therefore the set of rotations has a group structure, known as a rotation group.

In linear algebra, a circulant matrix is a square matrix in which all rows are composed of the same elements and each row is rotated one element to the right relative to the preceding row. It is a particular kind of Toeplitz matrix.

In mathematics, the discrete Laplace operator is an analog of the continuous Laplace operator, defined so that it has meaning on a graph or a discrete grid. For the case of a finite-dimensional graph, the discrete Laplace operator is more commonly called the Laplacian matrix.

In the mathematical field of graph theory, the Laplacian matrix, also called the graph Laplacian, admittance matrix, Kirchhoff matrix or discrete Laplacian, is a matrix representation of a graph. Named after Pierre-Simon Laplace, the graph Laplacian matrix can be viewed as a matrix form of the negative discrete Laplace operator on a graph approximating the negative continuous Laplacian obtained by the finite difference method.

<span class="mw-page-title-main">Centrality</span> Degree of connectedness within a graph

In graph theory and network analysis, indicators of centrality assign numbers or rankings to nodes within a graph corresponding to their network position. Applications include identifying the most influential person(s) in a social network, key infrastructure nodes in the Internet or urban networks, super-spreaders of disease, and brain networks. Centrality concepts were first developed in social network analysis, and many of the terms used to measure centrality reflect their sociological origin.

The Rayleigh–Ritz method is a direct numerical method of approximating eigenvalues, originated in the context of solving physical boundary value problems and named after Lord Rayleigh and Walther Ritz.

In linear algebra, it is often important to know which vectors have their directions unchanged by a given linear transformation. An eigenvector or characteristic vector is such a vector. Thus an eigenvector of a linear transformation is scaled by a constant factor when the linear transformation is applied to it: . The corresponding eigenvalue, characteristic value, or characteristic root is the multiplying factor .

In mathematics, preconditioning is the application of a transformation, called the preconditioner, that conditions a given problem into a form that is more suitable for numerical solving methods. Preconditioning is typically related to reducing a condition number of the problem. The preconditioned problem is then usually solved by an iterative method.

The expander mixing lemma intuitively states that the edges of certain -regular graphs are evenly distributed throughout the graph. In particular, the number of edges between two vertex subsets and is always close to the expected number of edges between them in a random -regular graph, namely .

In linear algebra, eigendecomposition is the factorization of a matrix into a canonical form, whereby the matrix is represented in terms of its eigenvalues and eigenvectors. Only diagonalizable matrices can be factorized in this way. When the matrix being factorized is a normal or real symmetric matrix, the decomposition is called "spectral decomposition", derived from the spectral theorem.

In statistics, principal component regression (PCR) is a regression analysis technique that is based on principal component analysis (PCA). More specifically, PCR is used for estimating the unknown regression coefficients in a standard linear regression model.

In the mathematical field of linear algebra, an arrowhead matrix is a square matrix containing zeros in all entries except for the first row, first column, and main diagonal, these entries can be any number. In other words, the matrix has the form

References

  1. 1 2 Zaki, Mohammed J.; Meira, Wagner Jr. (2014). Data Mining and Analysis: Fundamental Concepts and Algorithms. Cambridge University Press. ISBN   9780521766333.
  2. M. E. J. Newman. "The mathematics of networks" (PDF). Retrieved 2006-11-09.{{cite journal}}: Cite journal requires |journal= (help)
  3. Christian F. A. Negre, Uriel N. Morzan, Heidi P. Hendrickson, Rhitankar Pal, George P. Lisi, J. Patrick Loria, Ivan Rivalta, Junming Ho, Victor S. Batista. (2018). "Eigenvector centrality for characterization of protein allosteric pathways". Proceedings of the National Academy of Sciences. 115 (52): E12201–E12208. arXiv: 1706.02327 . Bibcode:2018PNAS..11512201N. doi: 10.1073/pnas.1810452115 . PMC   6310864 . PMID   30530700.{{cite journal}}: CS1 maint: multiple names: authors list (link)
  4. 1 2 David Austin. "How Google Finds Your Needle in the Web's Haystack". AMS.
  5. M. E. J. Newman. "The mathematics of networks" (PDF). Retrieved 2006-11-09.{{cite journal}}: Cite journal requires |journal= (help)
  6. 1 2 Fletcher, Jack McKay and Wennekers, Thomas (2017). "From Structure to Activity: Using Centrality Measures to Predict Neuronal Activity". International Journal of Neural Systems. 28 (2): 1750013. doi: 10.1142/S0129065717500137 . hdl: 10026.1/9713 . PMID   28076982.{{cite journal}}: CS1 maint: multiple names: authors list (link)
  7. Edmund Landau (1895). "Zur relativen Wertbemessung der Turnierresultate". Deutsches Wochenschach (11): 366–369. doi:10.1007/978-1-4615-4819-5_23.
  8. Holme, Peter (15 April 2019). "Firsts in network science" . Retrieved 17 April 2019.
  9. Altman, Alon; Tennenholtz, Moshe (2005). "Ranking systems". Proceedings of the 6th ACM conference on Electronic commerce - EC '05. New York, New York, USA: ACM Press. pp. 1–8. doi:10.1145/1064009.1064010. ISBN   1-59593-049-3.
  10. Palacios-Huerta, Ignacio; Volij, Oscar (2004). "The Measurement of Intellectual Influence" (PDF). Econometrica. 72 (3). The Econometric Society: 963–977. doi:10.1111/j.1468-0262.2004.00519.x. hdl: 10419/80143 . ISSN   0012-9682.
  11. Solá, Luis; Romance, Miguel; Criado, Regino; Flores, Julio; García del Amo, Alejandro; Boccaletti, Stefano (2013). "Eigenvector centrality of nodes in multiplex networks". Chaos: An Interdisciplinary Journal of Nonlinear Science. 23 (3): 033131. arXiv: 1305.7445 . Bibcode:2013Chaos..23c3131S. doi:10.1063/1.4818544. ISSN   1054-1500. PMID   24089967. S2CID   14556381.
  12. De Domenico, Manlio; Solè-Ribalta, ALbert; Omodei, Elisa; Gòmez, Sergio; Arenas, Alex (2015). "Ranking in interconnected multilayer networks reveals versatile nodes". Nature Communications. 6: 6868. arXiv: 1305.7445 . doi:10.1063/1.4818544. ISSN   2041-1723. PMID   25904405. S2CID   14556381.
  13. Cruz, Cesi; Labonne, Julien; Querubin, Pablo (2017). "Politician Family Networks and Electoral Outcomes: Evidence from the Philippines". American Economic Review. 107 (10). University of Chicago Press: 3006–37. doi: 10.1257/aer.20150343 .
  14. Jackson, Matthew O. (2010-11-01). Social and Economic Networks. Princeton University Press. doi:10.2307/j.ctvcm4gh1. ISBN   978-1-4008-3399-3. JSTOR   j.ctvcm4gh1.
  15. Elliott, Matthew; Golub, Benjamin (2019). "A Network Approach to Public Goods". Journal of Political Economy. 127 (2). University of Chicago Press: 730–776. doi:10.1086/701032. ISSN   0022-3808. S2CID   158834906.