Highly Cited Researchers is a list published annually by Clarivate of academic authors whose publications have received particularly high numbers of citations in academic journals indexed by Web of Science . [1] These constitute approximately 0.1% of all scientific researchers. [2]
The list has been published yearly since 2001. [3]
In 2016, Clarivate sent a congratulatory email to many researchers who had not actually received the designation. It was not disclosed how many researchers the congratulations were erroneously sent to. [4]
Over 1,000 researchers were excluded from the 2023 list in total, largely for unethical citation practices and other examples of poor academic integrity. [5] Specifically, researchers affiliated with Saudi Arabian universities saw a sharp decline from the previous years' rankings in 2023. [6] In April 2023, El País revealed that Saudi universities had been paying researchers to add their institution as an affiliation. [7] After correction, the number of HCRs affiliated with Saudi institutions dropped by 30%. [8] Following the incident, Clarivate released a statement supporting integrity in researcher affiliations. [9]
From 2001 to 2016, the percentage of HCRs affiliated with institutions from mainland China increased, reflecting the increased spending on research conducted by Chinese colleges and universities, though during this time, the number of affiliations specifically to the Chinese Academy of Sciences decreased. [10]
The proportion of women in the HCR list has been studied, with results suggesting that research by women receives less citations, as they represented only 16% of the researchers on the list. [11] Research has also shown that in the fields of chemistry, computer science, engineering, mathematics, and physics and astronomy, where the biggest disparity occurs, women constitute 4–7% of the HCRs although they make up over 25% of authors in the field. [12] (For further information, see: Women in STEM fields and Sexism in academia )
As of their 2022 list, Clarivate uses "performance statistics" from data in the Web of Science. There are 21 specific fields, and one for interdisciplinary science—Clarivate creates a list of papers that are in the top 1% most highly cited in their field, [a] and admission to the HCR list is based on an author's number of papers in the top 1%. [13]
The resulting list of highly cited researchers can be manipulated by citation cartels. After it was found that obscure researchers had displaced eminent mathematicians in the list, Clarivate had to remove the whole field of mathematics from the list in 2023. [14]
Many institutions post a press release, blog post, or news story upon a faculty member being recognized in the program. [b]
The list of HCRs has been used to create a specific list of authoritative researchers; one study specifically used HCRs to show trends of retractions in Iran. [20]
As of 2019, the list is also used as a factor in determining the rank of a college in the Academic Ranking of World Universities , which uses the number of affiliated HCRs in its ranking criteria. [21]
A citation is a reference to a source. More precisely, a citation is an abbreviated alphanumeric expression embedded in the body of an intellectual work that denotes an entry in the bibliographic references section of the work for the purpose of acknowledging the relevance of the works of others to the topic of discussion at the spot where the citation appears.
A citation index is a kind of bibliographic index, an index of citations between publications, allowing the user to easily establish which later documents cite which earlier documents. A form of citation index is first found in 12th-century Hebrew religious literature. Legal citation indexes are found in the 18th century and were made popular by citators such as Shepard's Citations (1873). In 1961, Eugene Garfield's Institute for Scientific Information (ISI) introduced the first citation index for papers published in academic journals, first the Science Citation Index (SCI), and later the Social Sciences Citation Index (SSCI) and the Arts and Humanities Citation Index (AHCI). American Chemical Society converted its printed Chemical Abstract Service into internet-accessible SciFinder in 2008. The first automated citation indexing was done by CiteSeer in 1997 and was patented. Other sources for such data include Google Scholar, Microsoft Academic, Elsevier's Scopus, and the National Institutes of Health's iCite.
The Academic Ranking of World Universities (ARWU), also known as the Shanghai Ranking, is one of the annual publications of world university rankings. The league table was originally compiled and issued by Shanghai Jiao Tong University in 2003, making it the first global university ranking with multifarious indicators.
A book review is a form of literary criticism in which a book is merely described or analyzed based on content, style, and merit.
Scientometrics is a subfield of informetrics that studies quantitative aspects of scholarly literature. Major research issues include the measurement of the impact of research papers and academic journals, the understanding of scientific citations, and the use of such measurements in policy and management contexts. In practice there is a significant overlap between scientometrics and other scientific fields such as information systems, information science, science of science policy, sociology of science, and metascience. Critics have argued that overreliance on scientometrics has created a system of perverse incentives, producing a publish or perish environment that leads to low-quality research.
Citation analysis is the examination of the frequency, patterns, and graphs of citations in documents. It uses the directed graph of citations — links from one document to another document — to reveal properties of the documents. A typical aim would be to identify the most important documents in a collection. A classic example is that of the citations between academic articles and books. For another example, judges of law support their judgements by referring back to judgements made in earlier cases. An additional example is provided by patents which contain prior art, citation of earlier patents relevant to the current claim. The digitization of patent data and increasing computing power have led to a community of practice that uses these citation data to measure innovation attributes, trace knowledge flows, and map innovation networks.
Citation impact or citation rate is a measure of how many times an academic journal article or book or author is cited by other articles, books or authors. Citation counts are interpreted as measures of the impact or influence of academic work and have given rise to the field of bibliometrics or scientometrics, specializing in the study of patterns of academic impact through citation analysis. The importance of journals can be measured by the average citation rate, the ratio of number of citations to number articles published within a given time period and in a given index, such as the journal impact factor or the citescore. It is used by academic institutions in decisions about academic tenure, promotion and hiring, and hence also used by authors in deciding which journal to publish in. Citation-like measures are also used in other fields that do ranking, such as Google's PageRank algorithm, software metrics, college and university rankings, and business performance indicators.
The h-index is an author-level metric that measures both the productivity and citation impact of the publications, initially used for an individual scientist or scholar. The h-index correlates with success indicators such as winning the Nobel Prize, being accepted for research fellowships and holding positions at top universities. The index is based on the set of the scientist's most cited papers and the number of citations that they have received in other publications. The index has more recently been applied to the productivity and impact of a scholarly journal as well as a group of scientists, such as a department or university or country. The index was suggested in 2005 by Jorge E. Hirsch, a physicist at UC San Diego, as a tool for determining theoretical physicists' relative quality and is sometimes called the Hirsch index or Hirsch number.
The Science Citation Index Expanded is a citation index originally produced by the Institute for Scientific Information (ISI) and created by Eugene Garfield.
The Social Sciences Citation Index (SSCI) is a commercial citation index product of Clarivate Analytics. It was originally developed by the Institute for Scientific Information from the Science Citation Index. The Social Sciences Citation Index is a multidisciplinary index which indexes over 3,400 journals across 58 social science disciplines – 1985 to present, and it has 122 million cited references – 1900 to present. It also includes a range of 3,500 selected items from some of the world's finest scientific and technical journals. It has a range of useful search functions such as 'cited reference searching', searching by author, subject, or title. Whilst the Social Sciences Citation Index provides extensive support in bibliographic analytics and research, a number of academic scholars have expressed criticisms relating to ideological bias and its English-dominant publishing nature.
Scientometrics is a monthly peer-reviewed academic journal covering the field of scientometrics. It publishes original studies, short communications, review papers, letters to the editor, and book reviews. It is published by Akadémiai Kiadó and Springer Science+Business Media and was established in 1978. Its founder and first editor-in-chief was Tibor Braun.
ResearcherID is an identifying system for scientific authors. The system was introduced in January 2008 by Thomson Reuters Corporation.
Ofer Dekel is a computer science researcher in the Machine Learning Department of Microsoft Research. He obtained his PhD in Computer Science from the Hebrew University of Jerusalem and is an affiliate faculty at the Computer Science & Engineering department at the University of Washington.
The CWTS Leiden Ranking is an annual global university ranking based exclusively on bibliometric indicators. The rankings are compiled by the Centre for Science and Technology Studies at Leiden University in the Netherlands. The Clarivate Analytics bibliographic database Web of Science is used as the source of the publication and citation data.
Science of the Total Environment is a weekly international peer-reviewed scientific journal covering environmental science. It was established in 1972 and is published by Elsevier. The editors-in-chief are Damià Barceló, Jay Gan and Philip Hopke.
Author-level metrics are citation metrics that measure the bibliometric impact of individual authors, researchers, academics, and scholars. Many metrics have been developed that take into account varying numbers of factors.
Microsoft Academic was a free internet-based academic search engine for academic publications and literature, developed by Microsoft Research in 2016 as a successor of Microsoft Academic Search. Microsoft Academic was shut down in 2022. Both OpenAlex and The Lens claim to be successors to Microsoft Academic.
The Leiden Manifesto for research metrics (LM) is a list of "ten principles to guide research evaluation", published as a comment in Volume 520, Issue 7548 of Nature, on 22 April 2015. It was formulated by public policy professor Diana Hicks, scientometrics professor Paul Wouters, and their colleagues at the 19th International Conference on Science and Technology Indicators, held between 3–5 September 2014 in Leiden, The Netherlands.
The science-wide author databases of standardized citation indicators is a multidimensional ranking of the world's scientists produced since 2015 by a team of researchers led by John P. A. Ioannidis at Stanford.