PageRank

Last updated
An animation of the PageRank algorithm running on a small network of pages. The size of the nodes represents the perceived importance of the page, and arrows represent hyperlinks. Page rank animation.gif
An animation of the PageRank algorithm running on a small network of pages. The size of the nodes represents the perceived importance of the page, and arrows represent hyperlinks.
A simple illustration of the Pagerank algorithm. The percentage shows the perceived importance, and the arrows represent hyperlinks. PageRanks-Example.svg
A simple illustration of the Pagerank algorithm. The percentage shows the perceived importance, and the arrows represent hyperlinks.

PageRank (PR) is an algorithm used by Google Search to rank web pages in their search engine results. It is named after both the term "web page" and co-founder Larry Page. PageRank is a way of measuring the importance of website pages. According to Google:

Contents

PageRank works by counting the number and quality of links to a page to determine a rough estimate of how important the website is. The underlying assumption is that more important websites are likely to receive more links from other websites. [1]

Currently, PageRank is not the only algorithm used by Google to order search results, but it is the first algorithm that was used by the company, and it is the best known. [2] [3] As of September 24, 2019, all patents associated with PageRank have expired. [4]

Description

PageRank is a link analysis algorithm and it assigns a numerical weighting to each element of a hyperlinked set of documents, such as the World Wide Web, with the purpose of "measuring" its relative importance within the set. The algorithm may be applied to any collection of entities with reciprocal quotations and references. The numerical weight that it assigns to any given element E is referred to as the PageRank of E and denoted by

A PageRank results from a mathematical algorithm based on the webgraph, created by all World Wide Web pages as nodes and hyperlinks as edges, taking into consideration authority hubs such as cnn.com or mayoclinic.org. The rank value indicates an importance of a particular page. A hyperlink to a page counts as a vote of support. The PageRank of a page is defined recursively and depends on the number and PageRank metric of all pages that link to it ("incoming links"). A page that is linked to by many pages with high PageRank receives a high rank itself.

Numerous academic papers concerning PageRank have been published since Page and Brin's original paper. [5] In practice, the PageRank concept may be vulnerable to manipulation. Research has been conducted into identifying falsely influenced PageRank rankings. The goal is to find an effective means of ignoring links from documents with falsely influenced PageRank. [6]

Other link-based ranking algorithms for Web pages include the HITS algorithm invented by Jon Kleinberg (used by Teoma and now Ask.com), the IBM CLEVER project, the TrustRank algorithm, the Hummingbird algorithm, [7] and the SALSA algorithm. [8]

History

The eigenvalue problem behind PageRank's algorithm was independently rediscovered and reused in many scoring problems. In 1895, Edmund Landau suggested using it for determining the winner of a chess tournament. [9] [10] The eigenvalue problem was also suggested in 1976 by Gabriel Pinski and Francis Narin, who worked on scientometrics ranking scientific journals, [11] in 1977 by Thomas Saaty in his concept of Analytic Hierarchy Process which weighted alternative choices, [12] and in 1995 by Bradley Love and Steven Sloman as a cognitive model for concepts, the centrality algorithm. [13] [14]

A search engine called "RankDex" from IDD Information Services, designed by Robin Li in 1996, developed a strategy for site-scoring and page-ranking. [15] Li referred to his search mechanism as "link analysis," which involved ranking the popularity of a web site based on how many other sites had linked to it. [16] RankDex, the first search engine with page-ranking and site-scoring algorithms, was launched in 1996. [17] Li filed a patent for the technology in RankDex in 1997; it was granted in 1999. [18] He later used it when he founded Baidu in China in 2000. [19] [20] Google founder Larry Page referenced Li's work as a citation in some of his U.S. patents for PageRank. [21] [17] [22]

Larry Page and Sergey Brin developed PageRank at Stanford University in 1996 as part of a research project about a new kind of search engine. An interview with Héctor García-Molina, Stanford Computer Science professor and advisor to Sergey, [23] provides background into the development of the page-rank algorithm. [24] Sergey Brin had the idea that information on the web could be ordered in a hierarchy by "link popularity": a page ranks higher as there are more links to it. [25] The system was developed with the help of Scott Hassan and Alan Steremberg, both of whom were cited by Page and Brin as being critical to the development of Google. [5] Rajeev Motwani and Terry Winograd co-authored with Page and Brin the first paper about the project, describing PageRank and the initial prototype of the Google search engine, published in 1998. [5] Shortly after, Page and Brin founded Google Inc., the company behind the Google search engine. While just one of many factors that determine the ranking of Google search results, PageRank continues to provide the basis for all of Google's web-search tools. [26]

The name "PageRank" plays on the name of developer Larry Page, as well as of the concept of a web page. [27] [28] The word is a trademark of Google, and the PageRank process has been patented ( U.S. patent 6,285,999 ). However, the patent is assigned to Stanford University and not to Google. Google has exclusive license rights on the patent from Stanford University. The university received 1.8 million shares of Google in exchange for use of the patent; it sold the shares in 2005 for US$336 million. [29] [30]

PageRank was influenced by citation analysis, early developed by Eugene Garfield in the 1950s at the University of Pennsylvania, and by Hyper Search, developed by Massimo Marchiori at the University of Padua. In the same year PageRank was introduced (1998), Jon Kleinberg published his work on HITS. Google's founders cite Garfield, Marchiori, and Kleinberg in their original papers. [5] [31]

Algorithm

The PageRank algorithm outputs a probability distribution used to represent the likelihood that a person randomly clicking on links will arrive at any particular page. PageRank can be calculated for collections of documents of any size. It is assumed in several research papers that the distribution is evenly divided among all documents in the collection at the beginning of the computational process. The PageRank computations require several passes, called "iterations", through the collection to adjust approximate PageRank values to more closely reflect the theoretical true value.

A probability is expressed as a numeric value between 0 and 1. A 0.5 probability is commonly expressed as a "50% chance" of something happening. Hence, a document with a PageRank of 0.5 means there is a 50% chance that a person clicking on a random link will be directed to said document.

Simplified algorithm

Assume a small universe of four web pages: A, B, C, and D. Links from a page to itself are ignored. Multiple outbound links from one page to another page are treated as a single link. PageRank is initialized to the same value for all pages. In the original form of PageRank, the sum of PageRank over all pages was the total number of pages on the web at that time, so each page in this example would have an initial value of 1. However, later versions of PageRank, and the remainder of this section, assume a probability distribution between 0 and 1. Hence the initial value for each page in this example is 0.25.

The PageRank transferred from a given page to the targets of its outbound links upon the next iteration is divided equally among all outbound links.

If the only links in the system were from pages B, C, and D to A, each link would transfer 0.25 PageRank to A upon the next iteration, for a total of 0.75.

Suppose instead that page B had a link to pages C and A, page C had a link to page A, and page D had links to all three pages. Thus, upon the first iteration, page B would transfer half of its existing value (0.125) to page A and the other half (0.125) to page C. Page C would transfer all of its existing value (0.25) to the only page it links to, A. Since D had three outbound links, it would transfer one third of its existing value, or approximately 0.083, to A. At the completion of this iteration, page A will have a PageRank of approximately 0.458.

In other words, the PageRank conferred by an outbound link is equal to the document's own PageRank score divided by the number of outbound links L( ).

In the general case, the PageRank value for any page u can be expressed as:

,

i.e. the PageRank value for a page u is dependent on the PageRank values for each page v contained in the set Bu (the set containing all pages linking to page u), divided by the number L(v) of links from page v.

Damping factor

The PageRank theory holds that an imaginary surfer who is randomly clicking on links will eventually stop clicking. The probability, at any step, that the person will continue following links is a damping factor d. The probability that they instead jump to any random page is 1 - d. Various studies have tested different damping factors, but it is generally assumed that the damping factor will be set around 0.85. [5]

The damping factor is subtracted from 1 (and in some variations of the algorithm, the result is divided by the number of documents (N) in the collection) and this term is then added to the product of the damping factor and the sum of the incoming PageRank scores. That is,

So any page's PageRank is derived in large part from the PageRanks of other pages. The damping factor adjusts the derived value downward. The original paper, however, gave the following formula, which has led to some confusion:

The difference between them is that the PageRank values in the first formula sum to one, while in the second formula each PageRank is multiplied by N and the sum becomes N. A statement in Page and Brin's paper that "the sum of all PageRanks is one" [5] and claims by other Google employees [32] support the first variant of the formula above.

Page and Brin confused the two formulas in their most popular paper "The Anatomy of a Large-Scale Hypertextual Web Search Engine", where they mistakenly claimed that the latter formula formed a probability distribution over web pages. [5]

Google recalculates PageRank scores each time it crawls the Web and rebuilds its index. As Google increases the number of documents in its collection, the initial approximation of PageRank decreases for all documents.

The formula uses a model of a random surfer who reaches their target site after several clicks, then switches to a random page. The PageRank value of a page reflects the chance that the random surfer will land on that page by clicking on a link. It can be understood as a Markov chain in which the states are pages, and the transitions are the links between pages – all of which are all equally probable.

If a page has no links to other pages, it becomes a sink and therefore terminates the random surfing process. If the random surfer arrives at a sink page, it picks another URL at random and continues surfing again.

When calculating PageRank, pages with no outbound links are assumed to link out to all other pages in the collection. Their PageRank scores are therefore divided evenly among all other pages. In other words, to be fair with pages that are not sinks, these random transitions are added to all nodes in the Web. This residual probability, d, is usually set to 0.85, estimated from the frequency that an average surfer uses his or her browser's bookmark feature. So, the equation is as follows:

where are the pages under consideration, is the set of pages that link to , is the number of outbound links on page , and is the total number of pages.

The PageRank values are the entries of the dominant right eigenvector of the modified adjacency matrix rescaled so that each column adds up to one. This makes PageRank a particularly elegant metric: the eigenvector is

where R is the solution of the equation

where the adjacency function is the ratio between number of links outbound from page j to page i to the total number of outbound links of page j. The adjacency function is 0 if page does not link to , and normalized such that, for each j

,

i.e. the elements of each column sum up to 1, so the matrix is a stochastic matrix (for more details see the computation section below). Thus this is a variant of the eigenvector centrality measure used commonly in network analysis.

Because of the large eigengap of the modified adjacency matrix above, [33] the values of the PageRank eigenvector can be approximated to within a high degree of accuracy within only a few iterations.

Google's founders, in their original paper, [31] reported that the PageRank algorithm for a network consisting of 322 million links (in-edges and out-edges) converges to within a tolerable limit in 52 iterations. The convergence in a network of half the above size took approximately 45 iterations. Through this data, they concluded the algorithm can be scaled very well and that the scaling factor for extremely large networks would be roughly linear in , where n is the size of the network.

As a result of Markov theory, it can be shown that the PageRank of a page is the probability of arriving at that page after a large number of clicks. This happens to equal where is the expectation of the number of clicks (or random jumps) required to get from the page back to itself.

One main disadvantage of PageRank is that it favors older pages. A new page, even a very good one, will not have many links unless it is part of an existing site (a site being a densely connected set of pages, such as Wikipedia).

Several strategies have been proposed to accelerate the computation of PageRank. [34]

Various strategies to manipulate PageRank have been employed in concerted efforts to improve search results rankings and monetize advertising links. These strategies have severely impacted the reliability of the PageRank concept,[ citation needed ] which purports to determine which documents are actually highly valued by the Web community.

Since December 2007, when it started actively penalizing sites selling paid text links, Google has combatted link farms and other schemes designed to artificially inflate PageRank. How Google identifies link farms and other PageRank manipulation tools is among Google's trade secrets.

Computation

PageRank can be computed either iteratively or algebraically. The iterative method can be viewed as the power iteration method [35] [36] or the power method. The basic mathematical operations performed are identical.

Iterative

At , an initial probability distribution is assumed, usually

.

where N is the total number of pages, and is page i at time 0.

At each time step, the computation, as detailed above, yields

where d is the damping factor,

or in matrix notation

where and is the column vector of length containing only ones.

The matrix is defined as

i.e.,

,

where denotes the adjacency matrix of the graph and is the diagonal matrix with the outdegrees in the diagonal.

The probability calculation is made for each page at a time point, then repeated for the next time point. The computation ends when for some small

,

i.e., when convergence is assumed.

Power method

If the matrix is a transition probability, i.e., column-stochastic and is a probability distribution (i.e., , where is matrix of all ones), then equation ( 2 ) is equivalent to

Hence PageRank is the principal eigenvector of . A fast and easy way to compute this is using the power method: starting with an arbitrary vector , the operator is applied in succession, i.e.,

,

until

.

Note that in equation ( 3 ) the matrix on the right-hand side in the parenthesis can be interpreted as

,

where is an initial probability distribution. n the current case

.

Finally, if has columns with only zero values, they should be replaced with the initial probability vector . In other words,

,

where the matrix is defined as

,

with

In this case, the above two computations using only give the same PageRank if their results are normalized:

.

Implementation

Python

importnumpyasnpdefpagerank(M,d:float=0.85):"""PageRank algorithm with explicit number of iterations. Returns ranking of nodes (pages) in the adjacency matrix.    Parameters    ----------    M : numpy array        adjacency matrix where M_i,j represents the link from 'j' to 'i', such that for all 'j'        sum(i, M_i,j) = 1    d : float, optional        damping factor, by default 0.85    Returns    -------    numpy array        a vector of ranks such that v_i is the i-th rank from [0, 1],    """N=M.shape[1]w=np.ones(N)/NM_hat=d*Mv=M_hat@w+(1-d)while(np.linalg.norm(w-v)>=1e-10):w=vv=M_hat@w+(1-d)returnvM=np.array([[0,0,0,0],[0,0,0,0],[1,0.5,0,0],[0,0.5,1,0]])v=pagerank(M,0.85)

Variations

PageRank of an undirected graph

The PageRank of an undirected graph is statistically close to the degree distribution of the graph , [37] but they are generally not identical: If is the PageRank vector defined above, and is the degree distribution vector

where denotes the degree of vertex , and is the edge-set of the graph, then, with , [38] shows that:

that is, the PageRank of an undirected graph equals to the degree distribution vector if and only if the graph is regular, i.e., every vertex has the same degree.

Ranking objects of two kinds

A generalization of PageRank for the case of ranking two interacting groups of objects was described by Daugulis. [39] In applications it may be necessary to model systems having objects of two kinds where a weighted relation is defined on object pairs. This leads to considering bipartite graphs. For such graphs two related positive or nonnegative irreducible matrices corresponding to vertex partition sets can be defined. One can compute rankings of objects in both groups as eigenvectors corresponding to the maximal positive eigenvalues of these matrices. Normed eigenvectors exist and are unique by the Perron or Perron–Frobenius theorem. Example: consumers and products. The relation weight is the product consumption rate.

Distributed algorithm for PageRank computation

Sarma et al. describe two random walk-based distributed algorithms for computing PageRank of nodes in a network. [40] One algorithm takes rounds with high probability on any graph (directed or undirected), where n is the network size and is the reset probability (, which is called the damping factor) used in the PageRank computation. They also present a faster algorithm that takes rounds in undirected graphs. In both algorithms, each node processes and sends a number of bits per round that are polylogarithmic in n, the network size.

Google Toolbar

The Google Toolbar long had a PageRank feature which displayed a visited page's PageRank as a whole number between 0 (least popular) and 10 (most popular). Google had not disclosed the specific method for determining a Toolbar PageRank value, which was to be considered only a rough indication of the value of a website. The "Toolbar Pagerank" was available for verified site maintainers through the Google Webmaster Tools interface. However, on October 15, 2009, a Google employee confirmed that the company had removed PageRank from its Webmaster Tools section, saying that "We've been telling people for a long time that they shouldn't focus on PageRank so much. Many site owners seem to think it's the most important metric for them to track, which is simply not true." [41]

The "Toolbar Pagerank" was updated very infrequently. It was last updated in November 2013. In October 2014 Matt Cutts announced that another visible pagerank update would not be coming. [42] In March 2016 Google announced it would no longer support this feature, and the underlying API would soon cease to operate. [43] On April 15, 2016, Google turned off display of PageRank Data in Google Toolbar, [44] though the PageRank continued to be used internally to rank content in search results. [45]

SERP rank

The search engine results page (SERP) is the actual result returned by a search engine in response to a keyword query. The SERP consists of a list of links to web pages with associated text snippets, paid ads, featured snippets, and Q&A. The SERP rank of a web page refers to the placement of the corresponding link on the SERP, where higher placement means higher SERP rank. The SERP rank of a web page is a function not only of its PageRank, but of a relatively large and continuously adjusted set of factors (over 200). [46] [ unreliable source? ] Search engine optimization (SEO) is aimed at influencing the SERP rank for a website or a set of web pages.

Positioning of a webpage on Google SERPs for a keyword depends on relevance and reputation, also known as authority and popularity. PageRank is Google's indication of its assessment of the reputation of a webpage: It is non-keyword specific. Google uses a combination of webpage and website authority to determine the overall authority of a webpage competing for a keyword. [47] The PageRank of the HomePage of a website is the best indication Google offers for website authority. [48]

After the introduction of Google Places into the mainstream organic SERP, numerous other factors in addition to PageRank affect ranking a business in Local Business Results. [49] When Google elaborated on the reasons for PageRank deprecation at Q&A #March 2016, they announced Links and Content as the Top Ranking Factors. RankBrain had earlier in October 2015 been announced as the #3 Ranking Factor, so the Top 3 Factors have been confirmed officially by Google. [50]

Google directory PageRank

The Google Directory PageRank was an 8-unit measurement. Unlike the Google Toolbar, which shows a numeric PageRank value upon mouseover of the green bar, the Google Directory only displayed the bar, never the numeric values. Google Directory was closed on July 20, 2011. [51]

False or spoofed PageRank

It was known that the PageRank shown in the Toolbar could easily be spoofed. Redirection from one page to another, either via a HTTP 302 response or a "Refresh" meta tag, caused the source page to acquire the PageRank of the destination page. Hence, a new page with PR 0 and no incoming links could have acquired PR 10 by redirecting to the Google home page. Spoofing can usually be detected by performing a Google search for a source URL; if the URL of an entirely different site is displayed in the results, the latter URL may represent the destination of a redirection.

Manipulating PageRank

For search engine optimization purposes, some companies offer to sell high PageRank links to webmasters. [52] As links from higher-PR pages are believed to be more valuable, they tend to be more expensive. It can be an effective and viable marketing strategy to buy link advertisements on content pages of quality and relevant sites to drive traffic and increase a webmaster's link popularity. However, Google has publicly warned webmasters that if they are or were discovered to be selling links for the purpose of conferring PageRank and reputation, their links will be devalued (ignored in the calculation of other pages' PageRanks). The practice of buying and selling [53] is intensely debated across the Webmaster community. Google advised webmasters to use the nofollow HTML attribute value on paid links. According to Matt Cutts, Google is concerned about webmasters who try to game the system, and thereby reduce the quality and relevance of Google search results. [52]

In 2019, Google offered a new type of tags that do not pass PageRank and thus do not have value for SEO link manipulation: rel="ugc" as a tag for user-generated content, such as comments; and rel="sponsored" tag for advertisements or other types of sponsored content. [54]

Even though PageRank has become less important for SEO purposes, the existence of back-links from more popular websites continues to push a webpage higher up in search rankings. [55]

Directed Surfer Model

A more intelligent surfer that probabilistically hops from page to page depending on the content of the pages and query terms the surfer is looking for. This model is based on a query-dependent PageRank score of a page which as the name suggests is also a function of query. When given a multiple-term query, , the surfer selects a according to some probability distribution, , and uses that term to guide its behavior for a large number of steps. It then selects another term according to the distribution to determine its behavior, and so on. The resulting distribution over visited web pages is QD-PageRank. [56]

Other uses

The mathematics of PageRank are entirely general and apply to any graph or network in any domain. Thus, PageRank is now regularly used in bibliometrics, social and information network analysis, and for link prediction and recommendation. It is used for systems analysis of road networks, and in biology, chemistry, neuroscience, and physics. [57]

Scientific research and academia

PageRank has been used to quantify the scientific impact of researchers. The underlying citation and collaboration networks are used in conjunction with pagerank algorithm in order to come up with a ranking system for individual publications which propagates to individual authors. The new index known as pagerank-index (Pi) is demonstrated to be fairer compared to h-index in the context of many drawbacks exhibited by h-index. [58]

For the analysis of protein networks in biology PageRank is also a useful tool. [59] [60]

In any ecosystem, a modified version of PageRank may be used to determine species that are essential to the continuing health of the environment. [61]

A similar newer use of PageRank is to rank academic doctoral programs based on their records of placing their graduates in faculty positions. In PageRank terms, academic departments link to each other by hiring their faculty from each other (and from themselves). [62]

A version of PageRank has recently been proposed as a replacement for the traditional Institute for Scientific Information (ISI) impact factor, [63] and implemented at Eigenfactor as well as at SCImago. Instead of merely counting total citations to a journal, the "importance" of each citation is determined in a PageRank fashion.

In neuroscience, the PageRank of a neuron in a neural network has been found to correlate with its relative firing rate. [64]

Internet use

Personalized PageRank is used by Twitter to present users with other accounts they may wish to follow. [65]

Swiftype's site search product builds a "PageRank that's specific to individual websites" by looking at each website's signals of importance and prioritizing content based on factors such as number of links from the home page. [66]

A Web crawler may use PageRank as one of a number of importance metrics it uses to determine which URL to visit during a crawl of the web. One of the early working papers [67] that were used in the creation of Google is Efficient crawling through URL ordering, [68] which discusses the use of a number of different importance metrics to determine how deeply, and how much of a site Google will crawl. PageRank is presented as one of a number of these importance metrics, though there are others listed such as the number of inbound and outbound links for a URL, and the distance from the root directory on a site to the URL.

The PageRank may also be used as a methodology to measure the apparent impact of a community like the Blogosphere on the overall Web itself. This approach uses therefore the PageRank to measure the distribution of attention in reflection of the Scale-free network paradigm.[ citation needed ]

Other applications

In 2005, in a pilot study in Pakistan, Structural Deep Democracy, SD2 [69] [70] was used for leadership selection in a sustainable agriculture group called Contact Youth. SD2 uses PageRank for the processing of the transitive proxy votes, with the additional constraints of mandating at least two initial proxies per voter, and all voters are proxy candidates. More complex variants can be built on top of SD2, such as adding specialist proxies and direct votes for specific issues, but SD2 as the underlying umbrella system, mandates that generalist proxies should always be used.

In sport the PageRank algorithm has been used to rank the performance of: teams in the National Football League (NFL) in the USA; [71] individual soccer players; [72] and athletes in the Diamond League. [73]

PageRank has been used to rank spaces or streets to predict how many people (pedestrians or vehicles) come to the individual spaces or streets. [74] [75] In lexical semantics it has been used to perform Word Sense Disambiguation, [76] Semantic similarity, [77] and also to automatically rank WordNet synsets according to how strongly they possess a given semantic property, such as positivity or negativity. [78]

How a traffic system changes its operational mode can be described by transitions between quasi-stationary states in correlation structures of traffic flow. PageRank has been used to identify and explore the dominant states among these quasi-stationary states in traffic systems. [79]

nofollow

In early 2005, Google implemented a new value, "nofollow", [80] for the rel attribute of HTML link and anchor elements, so that website developers and bloggers can make links that Google will not consider for the purposes of PageRank—they are links that no longer constitute a "vote" in the PageRank system. The nofollow relationship was added in an attempt to help combat spamdexing.

As an example, people could previously create many message-board posts with links to their website to artificially inflate their PageRank. With the nofollow value, message-board administrators can modify their code to automatically insert "rel='nofollow'" to all hyperlinks in posts, thus preventing PageRank from being affected by those particular posts. This method of avoidance, however, also has various drawbacks, such as reducing the link value of legitimate comments. (See: Spam in blogs#nofollow)

In an effort to manually control the flow of PageRank among pages within a website, many webmasters practice what is known as PageRank Sculpting [81] —which is the act of strategically placing the nofollow attribute on certain internal links of a website in order to funnel PageRank towards those pages the webmaster deemed most important. This tactic had been used since the inception of the nofollow attribute, but may no longer be effective since Google announced that blocking PageRank transfer with nofollow does not redirect that PageRank to other links. [82]

See also

Related Research Articles

<span class="mw-page-title-main">Discrete Fourier transform</span> Type of Fourier transform in discrete mathematics

In mathematics, the discrete Fourier transform (DFT) converts a finite sequence of equally-spaced samples of a function into a same-length sequence of equally-spaced samples of the discrete-time Fourier transform (DTFT), which is a complex-valued function of frequency. The interval at which the DTFT is sampled is the reciprocal of the duration of the input sequence. An inverse DFT (IDFT) is a Fourier series, using the DTFT samples as coefficients of complex sinusoids at the corresponding DTFT frequencies. It has the same sample-values as the original input sequence. The DFT is therefore said to be a frequency domain representation of the original input sequence. If the original sequence spans all the non-zero values of a function, its DTFT is continuous, and the DFT provides discrete samples of one cycle. If the original sequence is one cycle of a periodic function, the DFT provides all the non-zero values of one DTFT cycle.

A likelihood function measures how well a statistical model explains observed data by calculating the probability of seeing that data under different parameter values of the model. It is constructed from the joint probability distribution of the random variable that (presumably) generated the observations. When evaluated on the actual data points, it becomes a function solely of the model parameters.

<span class="mw-page-title-main">Multivariate normal distribution</span> Generalization of the one-dimensional normal distribution to higher dimensions

In probability theory and statistics, the multivariate normal distribution, multivariate Gaussian distribution, or joint normal distribution is a generalization of the one-dimensional (univariate) normal distribution to higher dimensions. One definition is that a random vector is said to be k-variate normally distributed if every linear combination of its k components has a univariate normal distribution. Its importance derives mainly from the multivariate central limit theorem. The multivariate normal distribution is often used to describe, at least approximately, any set of (possibly) correlated real-valued random variables, each of which clusters around a mean value.

<span class="mw-page-title-main">Noether's theorem</span> Statement relating differentiable symmetries to conserved quantities

Noether's theorem states that every continuous symmetry of the action of a physical system with conservative forces has a corresponding conservation law. This is the first of two theorems published by mathematician Emmy Noether in 1918. The action of a physical system is the integral over time of a Lagrangian function, from which the system's behavior can be determined by the principle of least action. This theorem only applies to continuous and smooth symmetries of physical space.

In mathematical optimization, the method of Lagrange multipliers is a strategy for finding the local maxima and minima of a function subject to equation constraints. It is named after the mathematician Joseph-Louis Lagrange.

<span class="mw-page-title-main">Hamiltonian mechanics</span> Formulation of classical mechanics using momenta

In physics, Hamiltonian mechanics is a reformulation of Lagrangian mechanics that emerged in 1833. Introduced by Sir William Rowan Hamilton, Hamiltonian mechanics replaces (generalized) velocities used in Lagrangian mechanics with (generalized) momenta. Both theories provide interpretations of classical mechanics and describe the same physical phenomena.

In linear algebra, an invertible matrix is a square matrix which has an inverse. In other words, if some other matrix is multiplied by the invertible matrix, the result can be multiplied by an inverse to undo the operation. Invertible matrices are the same size as their inverse.

<span class="mw-page-title-main">Logistic regression</span> Statistical model for a binary dependent variable

In statistics, the logistic model is a statistical model that models the log-odds of an event as a linear combination of one or more independent variables. In regression analysis, logistic regression estimates the parameters of a logistic model. In binary logistic regression there is a single binary dependent variable, coded by an indicator variable, where the two values are labeled "0" and "1", while the independent variables can each be a binary variable or a continuous variable. The corresponding probability of the value labeled "1" can vary between 0 and 1, hence the labeling; the function that converts log-odds to probability is the logistic function, hence the name. The unit of measurement for the log-odds scale is called a logit, from logistic unit, hence the alternative names. See § Background and § Definition for formal mathematics, and § Example for a worked example.

In probability theory and statistics, a Gaussian process is a stochastic process, such that every finite collection of those random variables has a multivariate normal distribution. The distribution of a Gaussian process is the joint distribution of all those random variables, and as such, it is a distribution over functions with a continuous domain, e.g. time or space.

In statistics, the matrix normal distribution or matrix Gaussian distribution is a probability distribution that is a generalization of the multivariate normal distribution to matrix-valued random variables.

Variational Bayesian methods are a family of techniques for approximating intractable integrals arising in Bayesian inference and machine learning. They are typically used in complex statistical models consisting of observed variables as well as unknown parameters and latent variables, with various sorts of relationships among the three types of random variables, as might be described by a graphical model. As typical in Bayesian inference, the parameters and latent variables are grouped together as "unobserved variables". Variational Bayesian methods are primarily used for two purposes:

  1. To provide an analytical approximation to the posterior probability of the unobserved variables, in order to do statistical inference over these variables.
  2. To derive a lower bound for the marginal likelihood of the observed data. This is typically used for performing model selection, the general idea being that a higher marginal likelihood for a given model indicates a better fit of the data by that model and hence a greater probability that the model in question was the one that generated the data.

In information theory, the cross-entropy between two probability distributions and , over the same underlying set of events, measures the average number of bits needed to identify an event drawn from the set when the coding scheme used for the set is optimized for an estimated probability distribution , rather than the true distribution .

<span class="mw-page-title-main">Scoring rule</span> Measure for evaluating probabilistic forecasts

In decision theory, a scoring rule provides evaluation metrics for probabilistic predictions or forecasts. While "regular" loss functions assign a goodness-of-fit score to a predicted value and an observed value, scoring rules assign such a score to a predicted probability distribution and an observed value. On the other hand, a scoring function provides a summary measure for the evaluation of point predictions, i.e. one predicts a property or functional , like the expectation or the median.

In statistics, the Kendall rank correlation coefficient, commonly referred to as Kendall's τ coefficient, is a statistic used to measure the ordinal association between two measured quantities. A τ test is a non-parametric hypothesis test for statistical dependence based on the τ coefficient. It is a measure of rank correlation: the similarity of the orderings of the data when ranked by each of the quantities. It is named after Maurice Kendall, who developed it in 1938, though Gustav Fechner had proposed a similar measure in the context of time series in 1897.

<span class="mw-page-title-main">Z-channel (information theory)</span>

In coding theory and information theory, a Z-channel or binary asymmetric channel is a communications channel used to model the behaviour of some data storage systems.

In probability theory and statistics, the Dirichlet-multinomial distribution is a family of discrete multivariate probability distributions on a finite support of non-negative integers. It is also called the Dirichlet compound multinomial distribution (DCM) or multivariate Pólya distribution. It is a compound probability distribution, where a probability vector p is drawn from a Dirichlet distribution with parameter vector , and an observation drawn from a multinomial distribution with probability vector p and number of trials n. The Dirichlet parameter vector captures the prior belief about the situation and can be seen as a pseudocount: observations of each outcome that occur before the actual data is collected. The compounding corresponds to a Pólya urn scheme. It is frequently encountered in Bayesian statistics, machine learning, empirical Bayes methods and classical statistics as an overdispersed multinomial distribution.

Ranking of query is one of the fundamental problems in information retrieval (IR), the scientific/engineering discipline behind search engines. Given a query q and a collection D of documents that match the query, the problem is to rank, that is, sort, the documents in D according to some criterion so that the "best" results appear early in the result list displayed to the user. Ranking in terms of information retrieval is an important concept in computer science and is used in many different applications such as search engine queries and recommender systems. A majority of search engines use ranking algorithms to provide users with accurate and relevant results.

<span class="mw-page-title-main">Google matrix</span> Stochastic matrix representing links between entities

A Google matrix is a particular stochastic matrix that is used by Google's PageRank algorithm. The matrix represents a graph with edges representing links between pages. The PageRank of each page can then be generated iteratively from the Google matrix using the power method. However, in order for the power method to converge, the matrix must be stochastic, irreducible and aperiodic.

<span class="mw-page-title-main">Poisson distribution</span> Discrete probability distribution

In probability theory and statistics, the Poisson distribution is a discrete probability distribution that expresses the probability of a given number of events occurring in a fixed interval of time if these events occur with a known constant mean rate and independently of the time since the last event. It can also be used for the number of events in other types of intervals than time, and in dimension greater than 1.

In computer science and data mining, MinHash is a technique for quickly estimating how similar two sets are. The scheme was invented by Andrei Broder, and initially used in the AltaVista search engine to detect duplicate web pages and eliminate them from search results. It has also been applied in large-scale clustering problems, such as clustering documents by the similarity of their sets of words.

References

Citations

  1. "Facts about Google and Competition". Archived from the original on 4 November 2011. Retrieved 12 July 2014.
  2. Sullivan, Danny (2007-04-26). "What Is Google PageRank? A Guide For Searchers & Webmasters". Search Engine Land. Archived from the original on 2016-07-03.
  3. Cutts, Matt. "Algorithms Rank Relevant Results Higher". Archived from the original on July 2, 2013. Retrieved 19 October 2015.
  4. "US7058628B1 - Method for node ranking in a linked database - Google Patents". Google Patents . Archived from the original on January 16, 2020. Retrieved September 14, 2019.
  5. 1 2 3 4 5 6 7 Brin, S.; Page, L. (1998). "The anatomy of a large-scale hypertextual Web search engine" (PDF). Computer Networks and ISDN Systems. 30 (1–7): 107–117. CiteSeerX   10.1.1.115.5930 . doi:10.1016/S0169-7552(98)00110-X. ISSN   0169-7552. S2CID   7587743. Archived (PDF) from the original on 2015-09-27.
  6. Gyöngyi, Zoltán; Berkhin, Pavel; Garcia-Molina, Hector; Pedersen, Jan (2006), "Link spam detection based on mass estimation", Proceedings of the 32nd International Conference on Very Large Data Bases (VLDB '06, Seoul, Korea) (PDF), pp. 439–450, archived (PDF) from the original on 2014-12-03.
  7. "FAQ: All About The New Google "Hummingbird" Algorithm". Search Engine Land. 26 September 2013. Archived from the original on 23 December 2018. Retrieved 18 December 2018.
  8. Wang, Ziyang. "Improved Link-Based Algorithms for Ranking Web Pages" (PDF). cs.nyu.edu. New York University, Department of Computer Science. Retrieved 7 August 2023.
  9. Landau, Edmund (1895). "Zur relativen Wertbemessung der Turnierresultate". Deutsches Wochenschach. 11 (42): 51–54.
  10. Sinn, Rainer; Ziegler, Günter M. (2022-10-31). "Landau on Chess Tournaments and Google's PageRank". arXiv: 2210.17300 [math.HO].
  11. Gabriel Pinski & Francis Narin (1976). "Citation influence for journal aggregates of scientific publications: Theory, with application to the literature of physics". Information Processing & Management. 12 (5): 297–312. doi:10.1016/0306-4573(76)90048-0.
  12. Thomas Saaty (1977). "A scaling method for priorities in hierarchical structures". Journal of Mathematical Psychology. 15 (3): 234–281. doi:10.1016/0022-2496(77)90033-5. hdl: 10338.dmlcz/101787 .
  13. Bradley C. Love & Steven A. Sloman. "Mutability and the determinants of conceptual transformability" (PDF). Proceedings of the Seventeenth Annual Conference of the Cognitive Science Society. pp. 654–659. Archived (PDF) from the original on 2017-12-23. Retrieved 2017-12-23.
  14. "How a CogSci undergrad invented PageRank three years before Google". bradlove.org. Archived from the original on 2017-12-11. Retrieved 2017-12-23.
  15. Li, Yanhong (August 6, 2002). "Toward a qualitative search engine". IEEE Internet Computing. 2 (4): 24–29. doi:10.1109/4236.707687.
  16. "The Rise of Baidu (That's Chinese for Google)". The New York Times . 17 September 2006. Archived from the original on 27 June 2019. Retrieved 16 June 2019.
  17. 1 2 "About: RankDex" Archived 2015-05-25 at the Wayback Machine , RankDex; accessed 3 May 2014.
  18. USPTO, "Hypertext Document Retrieval System and Method" Archived 2011-12-05 at the Wayback Machine , U.S. Patent number: 5920859, Inventor: Yanhong Li, Filing date: Feb 5, 1997, Issue date: Jul 6, 1999
  19. Greenberg, Andy, "The Man Who's Beating Google" Archived 2013-03-08 at the Wayback Machine , Forbes magazine, October 05, 2009
  20. "About: RankDex" Archived 2012-01-20 at the Wayback Machine , rankdex.com
  21. "Method for node ranking in a linked database". Google Patents. Archived from the original on 15 October 2015. Retrieved 19 October 2015.
  22. Altucher, James (March 18, 2011). "10 Unusual Things About Google". Forbes . Archived from the original on 16 June 2019. Retrieved 16 June 2019.
  23. Greg Wientjes. "Hector Garcia-Molina: Stanford Computer Science Professor and Advisor to Sergey". pp. minutes 25.45-32.50, 34.00–38.20. Retrieved 2019-12-06.
  24. Page, Larry, "PageRank: Bringing Order to the Web" (PDF). Archived (PDF) from the original on January 26, 2009. Retrieved 2022-10-06., Stanford Digital Library Project, talk. August 18, 1997 (archived 2002)
  25. 187-page study from Graz University, Austria Archived 2014-01-16 at the Wayback Machine , includes the note that also human brains are used when determining the page rank in Google.
  26. "Our products and services". Archived from the original on 2008-06-23. Retrieved 2011-05-27.
  27. David Vise & Mark Malseed (2005). The Google Story . Delacorte Press. p.  37. ISBN   978-0-553-80457-7.
  28. "Google Press Center: Fun Facts". Archived from the original on 2001-07-15.
  29. Lisa M. Krieger (1 December 2005). "Stanford Earns $336 Million Off Google Stock". San Jose Mercury News. Archived from the original on 8 April 2009. Retrieved 2009-02-25 via cited by redOrbit.
  30. Richard Brandt. "Starting Up. How Google got its groove". Stanford magazine. Archived from the original on 2009-03-10. Retrieved 2009-02-25.
  31. 1 2 Page, Lawrence; Brin, Sergey; Motwani, Rajeev; Winograd, Terry (1999). The PageRank citation ranking: Bringing order to the Web (Report). Archived from the original on 2006-04-27., published as a technical report on January 29, 1998 PDF Archived 2011-08-18 at the Wayback Machine
  32. Matt Cutts's blog: Straight from Google: What You Need to Know Archived 2010-02-07 at the Wayback Machine , see page 15 of his slides.
  33. Taher Haveliwala & Sepandar Kamvar (March 2003). "The Second Eigenvalue of the Google Matrix" (PDF). Stanford University Technical Report: 7056. arXiv: math/0307056 . Bibcode:2003math......7056N. Archived (PDF) from the original on 2008-12-17.
  34. Gianna M. Del Corso; Antonio Gullí; Francesco Romani (2004). "Fast PageRank Computation Via a Sparse Linear System (Extended Abstract)". In Stefano Leonardi (ed.). Algorithms and Models for the Web-Graph: Third International Workshop, WAW 2004, Rome, Italy, October 16, 2004. Proceedings. pp. 118–130. CiteSeerX   10.1.1.58.9060 . doi:10.1007/978-3-540-30216-2_10. ISBN   978-3-540-23427-2.
  35. Arasu, A.; Novak, J.; Tomkins, A.; Tomlin, J. (2002). "PageRank computation and the structure of the web: Experiments and algorithms". Proceedings of the Eleventh International World Wide Web Conference, Poster Track. Brisbane, Australia. pp. 107–117. CiteSeerX   10.1.1.18.5264 .
  36. Massimo Franceschet (2010). "PageRank: Standing on the shoulders of giants". arXiv: 1002.2858 [cs.IR].
  37. Nicola Perra and Santo Fortunato; Fortunato (September 2008). "Spectral centrality measures in complex networks". Phys. Rev. E. 78 (3): 36107. arXiv: 0805.3322 . Bibcode:2008PhRvE..78c6107P. doi:10.1103/PhysRevE.78.036107. PMID   18851105. S2CID   1755112.
  38. Vince Grolmusz (2015). "A Note on the PageRank of Undirected Graphs". Information Processing Letters. 115 (6–8): 633–634. arXiv: 1205.1960 . doi:10.1016/j.ipl.2015.02.015. S2CID   9855132.
  39. Peteris Daugulis; Daugulis (2012). "A note on a generalization of eigenvector centrality for bipartite graphs and applications". Networks. 59 (2): 261–264. arXiv: 1610.01544 . doi:10.1002/net.20442. S2CID   1436859.
  40. Atish Das Sarma; Anisur Rahaman Molla; Gopal Pandurangan; Eli Upfal (2015). "Fast Distributed PageRank Computation". Theoretical Computer Science. 561: 113–121. arXiv: 1208.3071 . doi:10.1016/j.tcs.2014.04.003. S2CID   10284718.
  41. Susan Moskwa. "PageRank Distribution Removed From WMT". Archived from the original on October 17, 2009. Retrieved October 16, 2009.
  42. Bartleman, Wil (2014-10-12). "Google Page Rank Update is Not Coming". Managed Admin. Archived from the original on 2015-04-02. Retrieved 2014-10-12.
  43. Schwartz, Barry (March 8, 2016). "Google has confirmed it is removing Toolbar PageRank". Search Engine Land . Archived from the original on March 10, 2016.
  44. Schwartz, Barry (18 April 2016). "Google Toolbar PageRank officially goes dark". Search Engine Land. Archived from the original on 2016-04-21.
  45. Southern, Matt (2016-04-19). "Google PageRank Officially Shuts its Doors to the Public". Search Engine Journal. Archived from the original on 2017-04-13.
  46. Fishkin, Rand; Jeff Pollard (April 2, 2007). "Search Engine Ranking Factors - Version 2". seomoz.org. Archived from the original on May 7, 2009. Retrieved May 11, 2009.
  47. Dover, D. Search Engine Optimization Secrets Indianapolis. Wiley. 2011.
  48. Viniker, D. The Importance of Keyword Difficulty Screening for SEO. Ed. Schwartz, M. Digital Guidebook Volume 5. News Press. p 160–164.
  49. "Ranking of listings: Ranking - Google Places Help". Archived from the original on 2012-05-26. Retrieved 2011-05-27.
  50. Clark, Jack. "Google Turning Its Lucrative Web Search Over to AI Machines". Bloomberg. Archived from the original on 25 March 2016. Retrieved 26 March 2016.
  51. Search Engine Watch: Google Directory Has Been Shut Down July 25, 2011
  52. 1 2 "How to report paid links". mattcutts.com/blog. April 14, 2007. Archived from the original on May 28, 2007. Retrieved 2007-05-28.
  53. "Google Link Schemes" Archived 2020-05-21 at the Wayback Machine links
  54. "Evolving". Google Developers. Retrieved 2022-02-08.
  55. "So...You Think SEO Has Changed". 19 March 2014. Archived from the original on 31 March 2014.
  56. Matthew Richardson & Pedro Domingos, A. (2001). The Intelligent Surfer:Probabilistic Combination of Link and Content Information in PageRank (PDF). pp. 1441–1448. Archived (PDF) from the original on 2016-03-04.
  57. Gleich, David F. (January 2015). "PageRank Beyond the Web". SIAM Review. 57 (3): 321–363. arXiv: 1407.5107 . doi:10.1137/140976649. S2CID   8375649.
  58. Senanayake, Upul; Piraveenan, Mahendra; Zomaya, Albert (2015). "The Pagerank-Index: Going beyond Citation Counts in Quantifying Scientific Impact of Researchers". PLOS ONE. 10 (8): e0134794. Bibcode:2015PLoSO..1034794S. doi: 10.1371/journal.pone.0134794 . ISSN   1932-6203. PMC   4545754 . PMID   26288312.
  59. G. Ivan & V. Grolmusz (2011). "When the Web meets the cell: using personalized PageRank for analyzing protein interaction networks". Bioinformatics. 27 (3): 405–7. doi: 10.1093/bioinformatics/btq680 . PMID   21149343.
  60. D. Banky and G. Ivan and V. Grolmusz (2013). "Equal opportunity for low-degree network nodes: a PageRank-based method for protein target identification in metabolic graphs". PLOS ONE. 8 (1): 405–7. Bibcode:2013PLoSO...854204B. doi: 10.1371/journal.pone.0054204 . PMC   3558500 . PMID   23382878.
  61. Burns, Judith (2009-09-04). "Google trick tracks extinctions". BBC News. Archived from the original on 2011-05-12. Retrieved 2011-05-27.
  62. Benjamin M. Schmidt & Matthew M. Chingos (2007). "Ranking Doctoral Programs by Placement: A New Method" (PDF). PS: Political Science and Politics. 40 (July): 523–529. CiteSeerX   10.1.1.582.9402 . doi:10.1017/s1049096507070771. S2CID   6012229. Archived (PDF) from the original on 2015-02-13.
  63. Johan Bollen; Marko A. Rodriguez; Herbert Van de Sompel (December 2006). "MESUR: Usage-based metrics of scholarly impact". Proceedings of the 7th ACM/IEEE-CS joint conference on Digital libraries. New York: Association for Computing Machinery. arXiv: cs.GL/0601030 . Bibcode:2006cs........1030B. doi:10.1145/1255175.1255273. ISBN   978-1-59593-644-8. S2CID   3115544.
  64. Fletcher, Jack McKay; Wennekers, Thomas (2017). "From Structure to Activity: Using Centrality Measures to Predict Neuronal Activity". International Journal of Neural Systems. 28 (2): 1750013. doi: 10.1142/S0129065717500137 . hdl: 10026.1/9713 . PMID   28076982.
  65. Gupta, Pankaj; Goel, Ashish; Lin, Jimmy; Sharma, Aneesh; Wang, Dong; Zadeh, Reza (2013). "WTF: The Who to Follow Service at Twitter". Proceedings of the 22nd International Conference on World Wide Web. ACM. pp. 505–514. doi:10.1145/2488388.2488433. ISBN   978-1-4503-2035-1. S2CID   207205045 . Retrieved 11 December 2018.
  66. Ha, Anthony (2012-05-08). "Y Combinator-Backed Swiftype Builds Site Search That Doesn't Suck". TechCrunch. Archived from the original on 2014-07-06. Retrieved 2014-07-08.
  67. "Working Papers Concerning the Creation of Google". Google. Archived from the original on November 28, 2006. Retrieved November 29, 2006.
  68. Cho, J.; Garcia-Molina, H.; Page, L. (1998). "Efficient crawling through URL ordering". Proceedings of the Seventh Conference on World Wide Web. Archived from the original on 2008-06-03.
  69. "Yahoo! Groups". Groups.yahoo.com. Archived from the original on 2013-10-04. Retrieved 2013-10-02.
  70. "Autopoietic Information Systems in Modern Organizations". CiteSeerX   10.1.1.148.9274 .
  71. Zack, Laurie; Lamb, Ron; Ball, Sarah (2012-12-31). "An application of Google's PageRank to NFL rankings". Involve: A Journal of Mathematics. 5 (4): 463–471. doi: 10.2140/involve.2012.5.463 . ISSN   1944-4184.
  72. Peña, Javier López; Touchette, Hugo (2012-06-28). "A network theory analysis of football strategies". arXiv: 1206.6904 [math.CO].
  73. Beggs, Clive B.; Shepherd, Simon J.; Emmonds, Stacey; Jones, Ben (2017-06-02). Zhou, Wei-Xing (ed.). "A novel application of PageRank and user preference algorithms for assessing the relative performance of track athletes in competition". PLOS ONE. 12 (6): e0178458. Bibcode:2017PLoSO..1278458B. doi: 10.1371/journal.pone.0178458 . ISSN   1932-6203. PMC   5456068 . PMID   28575009.
  74. B. Jiang (2006). "Ranking spaces for predicting human movement in an urban environment". International Journal of Geographical Information Science. 23 (7): 823–837. arXiv: physics/0612011 . Bibcode:2009IJGIS..23..823J. doi:10.1080/13658810802022822. S2CID   26880621.
  75. Jiang B.; Zhao S. & Yin J. (2008). "Self-organized natural roads for predicting traffic flow: a sensitivity study". Journal of Statistical Mechanics: Theory and Experiment. P07008 (7): 008. arXiv: 0804.1630 . Bibcode:2008JSMTE..07..008J. doi:10.1088/1742-5468/2008/07/P07008. S2CID   118605727.
  76. Roberto Navigli, Mirella Lapata. "An Experimental Study of Graph Connectivity for Unsupervised Word Sense Disambiguation" Archived 2010-12-14 at the Wayback Machine . IEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI), 32(4), IEEE Press, 2010, pp. 678–692.
  77. M. T. Pilehvar, D. Jurgens and R. Navigli. Align, Disambiguate and Walk: A Unified Approach for Measuring Semantic Similarity. Archived 2013-10-01 at the Wayback Machine . Proc. of the 51st Annual Meeting of the Association for Computational Linguistics (ACL 2013), Sofia, Bulgaria, August 4–9, 2013, pp. 1341-1351.
  78. Andrea Esuli & Fabrizio Sebastiani. "PageRanking WordNet synsets: An Application to Opinion-Related Properties" (PDF). In Proceedings of the 35th Meeting of the Association for Computational Linguistics, Prague, CZ, 2007, pp. 424–431. Archived (PDF) from the original on June 28, 2007. Retrieved June 30, 2007.
  79. Wang S.; Schreckenberg M.; Guhr T (2023). "Transitions between quasi-stationary states in traffic systems: Cologne orbital motorways as an example". Journal of Statistical Mechanics: Theory and Experiment. 2023 (9): 093401. arXiv: 2302.14596 . Bibcode:2023JSMTE2023i3401W. doi: 10.1088/1742-5468/acf210 . S2CID   257232659.
  80. "Preventing Comment Spam". Google. Archived from the original on June 12, 2005. Retrieved January 1, 2005.
  81. "PageRank Sculpting: Parsing the Value and Potential Benefits of Sculpting PR with Nofollow". SEOmoz. 14 October 2008. Archived from the original on 2011-05-14. Retrieved 2011-05-27.
  82. "PageRank sculpting". Mattcutts.com. 2009-06-15. Archived from the original on 2011-05-11. Retrieved 2011-05-27.

Sources

Relevant patents