Google Penguin is a codename [1] for a Google algorithm update that was first announced on April 24, 2012. The update was aimed at decreasing search engine rankings of websites that violate Google's Webmaster Guidelines [2] by using now declared Grey Hat SEM techniques involved in increasing artificially the ranking of a webpage by manipulating the number of links pointing to the page. Such tactics are commonly described as link schemes. [3] According to Google's John Mueller, [1] as of 2013, Google announced all updates to the Penguin filter to the public. [4]
By Google's estimates, [5] Penguin affected approximately 3.1% of search queries in English, about 3% of queries in languages like German, Chinese, and Arabic, and an even greater percentage of them in "highly spammed" languages. On May 25, 2012, Google unveiled another Penguin update, called Penguin 1.1. This update, according to Matt Cutts, former head of webspam at Google, was supposed to affect less than one-tenth of a percent of English searches. The guiding principle for the update was to penalize websites that were using manipulative techniques to achieve high rankings. Pre-Penguin sites commonly used negative link building techniques to rank highly and get traffic. Once Penguin was rolled out, it meant that content was key, and those with great content would be recognised and those with little or spammy content would be penalised and receive no ranking benefits. [6] The purpose according to Google was to catch excessive spammers. Allegedly, few websites lost search rankings on Google for specific keywords during the Panda and Penguin rollouts. [7] Google specifically mentions that doorway pages, which are only built to attract search engine traffic, are against their webmaster guidelines.
In January 2012, the so-called Page Layout Algorithm Update [8] (also known as the Top Heavy Update) [9] was released, which targeted websites with too many ads, or too little content above the fold.
Penguin 3 was released October 5, 2012, and affected 0.3% of queries. [10] Penguin 4 (also known as Penguin 2.0) was released on May 22, 2013, and affected 2.3% of queries. [11] Penguin 5 (also known as Penguin 2.1) [12] was released on October 4, 2013, affected around 1% of queries, and has been the most recent of the Google Penguin algorithm updates. [13]
Google was reported to have released Penguin 3.0 on October 18, 2014. [14]
On October 21, 2014, Google's Pierre Farr confirmed that Penguin 3.0 was an algorithm "refresh", with no new signals added. [15]
On April 7, 2015, Google's John Mueller said in a Google+ hangout that both Penguin and Panda "currently are not updating the data regularly" and that updates must be pushed out manually. This confirms that the algorithm is not updated continuously which was believed to be the case earlier on in the year. [16]
The strategic goal that Panda, Penguin, and the page layout update share is to display higher quality websites at the top of Google's search results. However, sites that were downranked as the result of these updates have different sets of characteristics. The main target of Google Penguin is to focus on The so-called "black-hat" link-building strategies, such as link buying, link farming, automated links, PBNs, and others. [17]
In a Google+ Hangout on April 15, 2016, John Mueller said "I am pretty sure when we start rolling out [Penguin] we will have a message to kind of post but at the moment I don't have anything specific to kind of announce." [18]
On September 23, 2016 Google announced that Google Penguin was now part of the core algorithm [19] meaning that it updates in real time. Hence there will no longer be announcements by Google relating to future refreshes. [20] Real-time also means that websites are evaluated in real-time and rankings impacted in real-time. During the last years webmasters instead always had to wait for the roll-out of the next update to get out of a Penguin penalty. Also, Google Penguin 4.0 is more granular as opposed to previous updates, since it may affect a website on a URL-basis as opposed to always affecting a whole website. Finally, Penguin 4.0 [21] [22] differs from previous Penguin versions since it does not demote a web site when it finds bad links. Instead it discounts the links, meaning it ignores them and they no longer count toward the website's ranking. As a result of this, there is less need to use the disavow file. [21] Google uses both algorithm and human reviewers to identify links that are unnatural (artificial), manipulative or deceptive and includes these in its Manual Actions report for websites. [23]
Two days after the Penguin update was released Google prepared a feedback form, [24] designed for two categories of users: those who want to report web spam that still ranks highly after the search algorithm change, and those who think that their site got unfairly hit by the update. Google also has a reconsideration form through Google Webmaster Tools.
In January 2015, Google's John Mueller said that a Penguin penalty can be removed by simply building good links. The usual process is to remove bad links manually or by using Google's Disavow tool and then filing a reconsideration request. [25] Mueller elaborated on this by saying the algorithm looks at the percentage of good links versus bad links, so by building more good links it may tip the algorithm in your favor which would lead to recovery. [26]
Google Search is a search engine operated by Google. It allows users to search for information on the Web by entering keywords or phrases. Google Search uses algorithms to analyze and rank websites based on their relevance to the search query. It is the most popular search engine worldwide.
Spamdexing is the deliberate manipulation of search engine indexes. It involves a number of methods, such as link building and repeating related and/or unrelated phrases, to manipulate the relevance or prominence of resources indexed in a manner inconsistent with the purpose of the indexing system.
Search engine optimization (SEO) is the process of improving the quality and quantity of website traffic to a website or a web page from search engines. SEO targets unpaid traffic rather than direct traffic or paid traffic. Unpaid traffic may originate from different kinds of searches, including image search, video search, academic search, news search, and industry-specific vertical search engines.
On the World Wide Web, a link farm is any group of websites that all hyperlink to other sites in the group for the purpose of increasing SEO rankings. In graph theoretic terms, a link farm is a clique. Although some link farms can be created by hand, most are created through automated programs and services. A link farm is a form of spamming the index of a web search engine. Other link exchange systems are designed to allow individual websites to selectively exchange links with other relevant websites, and are not considered a form of spamdexing.
A metasearch engine is an online information retrieval tool that uses the data of a web search engine to produce its own results. Metasearch engines take input from a user and immediately query search engines for results. Sufficient data is gathered, ranked, and presented to the users.
The anchor text, link label, or link text is the visible, clickable text in an HTML hyperlink. The term "anchor" was used in older versions of the HTML specification for what is currently referred to as the "a element", or <a>
. The HTML specification does not have a specific term for anchor text, but refers to it as "text that the a element wraps around". In XML terms, the anchor text is the content of the element, provided that the content is text.
The sandbox effect is a theory about the way Google ranks web pages in its index. It is the subject of much debate—its existence has been written about since 2004, but not confirmed, with several statements to the contrary.
Matthew Cutts is an American software engineer. Cutts is the former Administrator of the United States Digital Service. He was first appointed as acting administrator, to later be confirmed as full administrator in October 2018. Cutts previously worked with Google as part of the search quality team on search engine optimization issues. He is the former head of the web spam team at Google.
An SEO contest is a prize activity that challenges search engine optimization (SEO) practitioners to achieve high ranking under major search engines such as Google, Yahoo, and MSN using certain keyword(s). This type of contest is controversial because it often leads to massive amounts of link spamming as participants try to boost the rankings of their pages by any means available. The SEO competitors hold the activity without the promotion of a product or service in mind, or they may organize a contest in order to market something on the Internet. Participants can showcase their skills and potentially discover and share new techniques for promoting websites.
nofollow is a setting on a web page hyperlink that directs search engines not to use the link for page ranking calculations. It is specified in the page as a type of link relation; that is: <a rel="nofollow" ...>
. Because search engines often calculate a site's importance according to the number of hyperlinks from other sites, the nofollow
setting allows website authors to indicate that the presence of a link is not an endorsement of the target site's importance.
Google Search Console is a web service by Google which allows webmasters to check indexing status, search queries, crawling errors and optimize visibility of their websites.
Mahalo.com was a web directory and Internet-based knowledge exchange launched in May 2007 by Jason Calacanis. It differentiated itself from algorithmic search engines like Google and Ask.com, as well as other directory sites like DMOZ and Yahoo! by tracking and building hand-crafted result sets for many of the currently popular search terms.
In the field of search engine optimization (SEO), link building describes actions aimed at increasing the number and quality of inbound links to a webpage with the goal of increasing the search engine rankings of that page or website. Briefly, link building is the process of establishing relevant hyperlinks to a website from external sites. Link building can increase the number of high-quality links pointing to a website, in turn increasing the likelihood of the website ranking highly in search engine results. Link building is also a proven marketing tactic for increasing brand awareness.
PageRank (PR) is an algorithm used by Google Search to rank web pages in their search engine results. It is named after both the term "web page" and co-founder Larry Page. PageRank is a way of measuring the importance of website pages. According to Google:
PageRank works by counting the number and quality of links to a page to determine a rough estimate of how important the website is. The underlying assumption is that more important websites are likely to receive more links from other websites.
A canonical link element is an HTML element that helps webmasters prevent duplicate content issues in search engine optimization by specifying the "canonical" or "preferred" version of a web page. It is described in RFC 6596, which went live in April 2012.
Yandex Search is a search engine owned by the company Yandex, based in Russia. In January 2015, Yandex Search generated 51.2% of all of the search traffic in Russia according to LiveInternet.
Google Panda is a major change to Google's search results ranking algorithm that was first released in February 2011. The change aimed to lower the rank of "low-quality sites" or "thin sites", in particular "content farms", and return higher-quality sites near the top of the search results.
Google Search, offered by Google, is the most widely used search engine on the World Wide Web as of 2023, with over eight billion searches a day. This page covers key events in the history of Google's search service.
This page provides a full timeline of web search engines, starting from the WHOis in 1982, the Archie search engine in 1990, and subsequent developments in the field. It is complementary to the history of web search engines page that provides more qualitative detail on the history.
Google Pigeon is the code name given to one of Google's local search algorithm updates. This update was released on July 24, 2014. It is aimed to increase the ranking of local listings in a search.