Author-level metrics

Last updated

Author-level metrics are citation metrics that measure the bibliometric impact of individual authors, researchers, academics, and scholars. Many metrics have been developed that take into account varying numbers of factors (from only considering the total number of citations, to looking at their distribution across papers or journals using statistical or graph-theoretic principles).

Contents

These quantitative comparisons between researchers are mostly done to distribute resources (such as money and academic positions). However, there is still debate in the academic world about how effectively author-level metrics accomplish this objective. [1] [2] [3]

Author-level metrics differ from journal-level metrics, which attempt to measure the bibliometric impact of academic journals rather than individuals, and from article-level metrics, which attempt to measure the impact of individual articles. However, metrics originally developed for academic journals can be reported at researcher level, such as the author-level eigenfactor [4] and the author impact factor. [5]

List of metrics

NameDescription
h-indexFormally, if f is the function that corresponds to the number of citations for each publication, the h-index is computed as follows. First, we order the values of f from the largest to the lowest value. Then, we look for the last position in which f is greater than or equal to the position (we call h this position). For example, if we have a researcher with 5 publications A, B, C, D, and E with 10, 8, 5, 4, and 3 citations, respectively, the h-index is equal to 4 because the 4th publication has 4 citations and the 5th has only 3. In contrast, if the same publications have 25, 8, 5, 3, and 3 citations, then the index is 3 because the fourth paper has only 3 citations. [1]
Individual h-indexAn individual h-index normalized by the number of authors has been proposed: , with being the number of authors considered in the papers. [6] It was found that the distribution of the h-index, although it depends on the field, can be normalized by a simple rescaling factor. For example, assuming as standard the hs for biology, the distribution of h for mathematics collapses with it if this h is multiplied by three, that is, a mathematician with h = 3 is equivalent to a biologist with h = 9. This method has not been readily adopted, perhaps because of its complexity.
Fractional h-indexTo avoid incentivizing hyperautorship with more than 100 coauthors per paper it might be simpler to divide citation counts by the number of authors before ordering the papers and obtaining the fractional h-index, as originally suggested by Hirsch. This index, also called h-frac, is not highly correlated with h-index and is currently correlated with scientific awards. [7]
h2Three additional metrics have been proposed: h2 lower, h2 center, and h2 upper, to give a more accurate representation of the distribution shape. The three h2 metrics measure the relative area within a scientist's citation distribution in the low impact area, h2 lower, the area captured by the h-index, h2 center, and the area from publications with the highest visibility, h2 upper. Scientists with high h2 upper percentages are perfectionists, whereas scientists with high h2 lower percentages are mass producers. As these metrics are percentages, they are intended to give a qualitative description to supplement the quantitative h-index. [8]
Field-weighted Citation ImpactField-weighted Citation Impact (FWCI) is an author-level metric introduced and applied by Scopus SciVal. [9] FWCI equals to the total citations actually received divided by the total citations that would be expected based on the average of the considered field. FWCI of 1 means that the output performs just as expected for the global average. More than 1 means that the author outperforms the average, and less than 1 means that the author underperforms. For instance, means % more likely to be cited. [10] [11]
Normalized h-indexThe h-index has been shown to have a strong discipline bias. However, a simple normalization by the average h of scholars in a discipline d is an effective way to mitigate this bias, obtaining a universal impact metric that allows comparison of scholars across different disciplines. [12]
Author-level EigenfactorAuthor-level Eigenfactor is a version of Eigenfactor for single authors. [13] Eigenfactor regards authors as nodes in a network of citations. The score of an author according to this metric is his or her eigenvector centrality in the network.
Erdős numberIt has been argued that "For an individual researcher, a measure such as Erdős number captures the structural properties of the network whereas the h-index captures the citation impact of the publications. One can be easily convinced that ranking in coauthorship networks should take into account both measures to generate a realistic and acceptable ranking." Several author ranking systems have been proposed already, for instance the Phys Author Rank Algorithm. [14]
i-10-indexThe i-10 index indicates the number of academic publications an author has written that have been cited by at least 10 sources. It was introduced in July 2011 by Google as part of their work on Google Scholar. [15]
RG ScoreResearchGate Score or RG Score is an author-level metric introduced by ResearchGate in 2012. [16] According to ResearchGate's CEO Dr. Ijad Madisch, “[t]he RG Score allows real-time feedback from the people who matter: the scientists themselves.” [17] RG Score has been reported to be correlated with existing author-level metrics and has an undisclosed calculation methodology. [18] [19] [20] [21] Two studies reported that RG Score seems to incorporate the journal impact factors into the calculation. [20] [21] The RG Score was found to be negatively correlated with network centrality – users that are the most active on ResearchGate usually do not have high RG scores. [22] It was also found to be strongly positively correlated with Quacquarelli Symonds university rankings at the institutional level, but only weakly with Elsevier SciVal rankings of individual authors. [23] While it was found to be correlated with different university rankings, the correlation in between these rankings themselves was higher. [18]
m-indexThe m-index is defined as h/n, where h is the h-index and n is the number of years since the first published paper of the scientist; [1] also called m-quotient. [24] [25]
g-indexFor g-index is introduced in 2006 as largest number of top articles, which have received together at least citations. [26]
e-indexThe e-index, the square root of surplus citations for the h-set beyond h2, complements the h-index for ignored citations, and therefore is especially useful for highly cited scientists and for comparing those with the same h-index (iso-h-index group). [27] [28]
c-indexThe c-index accounts not only for the citations but for the quality of the citations in terms of the collaboration distance between citing and cited authors. A scientist has c-index n if n of [his/her] N citations are from authors which are at collaboration distance at least n, and the other (Nn) citations are from authors which are at collaboration distance at most n. [29]
o-indexThe o-index corresponds to the geometric mean of the h-index and the most cited paper of a researcher. [30]
RA-indexThe RA-index accommodates improving the sensitivity of the h-index on the number of highly cited papers and has many cited paper and uncited paper under the h-core. This improvement can enhance the measurement sensitivity of the h-index. [31]
L-indexL-index combines the number of citations, the number of coauthors, the age of publications into a single value, which is independent of the number of publications and conveniently ranges from 0.0 to 9.9. [32] With c as number of citations, a as number of authors and y as number of years, L-index is defined by the formula:

L-index is automatically calculated by the Exaly database. [33]

s-indexAn s-index, accounting for the non-entropic distribution of citations, has been proposed and it has been shown to be in a very good correlation with h. [34]
w-indexw-index is defined as follow: if w of a researcher's papers have at least citations each and the other papers have fewer than citations, that researcher's w-index is w. [35]
Author Impact FactorAuthor Impact Factor (AIF) is the Impact Factor applied to authors. [5] The AIF of an author in year is the mean number of citations given by papers published in year to papers published by in a period of years before year . Unlike the h-index, AIF is able to capture trends and variations of the impact of the scientific output of scientists over time, which is a growing measure taking into account the whole career path.

Additional variations of h-index

There are a number of models proposed to incorporate the relative contribution of each author to a paper, for instance by accounting for the rank in the sequence of authors. [36] A generalization of the h-index and some other indices that gives additional information about the shape of the author's citation function (heavy-tailed, flat/peaked, etc.) has been proposed. [37] Because the h-index was never meant to measure future publication success, recently, a group of researchers has investigated the features that are most predictive of future h-index. It is possible to try the predictions using an online tool. [38] However, later work has shown that since h-index is a cumulative measure, it contains intrinsic auto-correlation that led to significant overestimation of its predictability. Thus, the true predictability of future h-index is much lower compared to what has been claimed before. [39] The h-index can be timed to analyze its evolution during one's career, employing different time windows. [40]

Criticism

Some academics, such as physicist Jorge E. Hirsch, have praised author-level metrics as a "useful yardstick with which to compare, in an unbiased way, different individuals competing for the same resource when an important evaluation criterion is scientific achievement." [1] However, other members of the scientific community, and even Hirsch himself, [41] have criticized them as particularly susceptible to gaming the system. [2] [3] [42]

Work in bibliometrics has demonstrated multiple techniques for the manipulation of popular author-level metrics. The most used metric h-index can be manipulated through self-citations, [43] [44] [45] and even computer-generated nonsense documents can be used for that purpose, for example using SCIgen. [46] Metrics can also be manipulated by coercive citation, a practice in which an editor of a journal forces authors to add spurious citations to their own articles before the journal will agree to publish it. [47] [48]

Additionally, if the h-index is considered as a decision criterion for research funding agencies, the game-theoretic solution to this competition implies increasing the average length of coauthors' lists. [49] A study analyzing >120 million papers in the specific field of biology showed that the validity of citation-based measures is being compromised and their usefulness is lessening. [50] As predicted by Goodhart's law, quantity of publications is not a good metric anymore as a result of shorter papers and longer author lists.

Leo Szilard, the inventor of the nuclear chain reaction, also expressed criticism of the decision-making system for scientific funding in his book "The Voice of the Dolphins and Other Stories". [51] Senator J. Lister Hill read excerpts of this criticism in a 1962 senate hearing on the slowing of government-funded cancer research. [52] Szilard's work focuses on metrics slowing scientific progress, rather than on specific methods of gaming:

"As a matter of fact, I think it would be quite easy. You could set up a foundation, with an annual endowment of thirty million dollars. Research workers in need of funds could apply for grants, if they could mail out a convincing case. Have ten committees, each committee, each composed of twelve scientists, appointed to pass on these applications. Take the most active scientists out of the laboratory and make them members of these committees. And the very best men in the field should be appointed as chairman at salaries of fifty thousand dollars each. Also have about twenty prizes of one hundred thousand dollars each for the best scientific papers of the year. This is just about all you would have to do. Your lawyers could easily prepare a charter for the foundation. As a matter of fact, any of the National Science Foundation bills which were introduced in the Seventy-ninth and Eightieth Congress could perfectly well serve as a model."

"First of all, the best scientists would be removed from their laboratories and kept busy on committees passing on applications for funds. Secondly the scientific workers in need of funds would concentrate on problems which were considered promising and were pretty certain to lead to publishable results. For a few years there might be a great increase in scientific output; but by going after the obvious, pretty soon science would dry out. Science would become something like a parlor game. Somethings would be considered interesting, others not. There would be fashions. Those who followed the fashions would get grants. Those who wouldn’t would not, and pretty soon they would learn to follow the fashion, too." [51]

See also

Related Research Articles

<span class="mw-page-title-main">Citation</span> Reference to a source

A citation is a reference to a source. More precisely, a citation is an abbreviated alphanumeric expression embedded in the body of an intellectual work that denotes an entry in the bibliographic references section of the work for the purpose of acknowledging the relevance of the works of others to the topic of discussion at the spot where the citation appears.

<span class="mw-page-title-main">Citation index</span> Index of citations between publications

A citation index is a kind of bibliographic index, an index of citations between publications, allowing the user to easily establish which later documents cite which earlier documents. A form of citation index is first found in 12th-century Hebrew religious literature. Legal citation indexes are found in the 18th century and were made popular by citators such as Shepard's Citations (1873). In 1961, Eugene Garfield's Institute for Scientific Information (ISI) introduced the first citation index for papers published in academic journals, first the Science Citation Index (SCI), and later the Social Sciences Citation Index (SSCI) and the Arts and Humanities Citation Index (AHCI). American Chemical Society converted its printed Chemical Abstract Service into internet-accessible SciFinder in 2008. The first automated citation indexing was done by CiteSeer in 1997 and was patented. Other sources for such data include Google Scholar, Microsoft Academic, Elsevier's Scopus, and the National Institutes of Health's iCite.

The impact factor (IF) or journal impact factor (JIF) of an academic journal is a scientometric index calculated by Clarivate that reflects the yearly mean number of citations of articles published in the last two years in a given journal, as indexed by Clarivate's Web of Science.

<span class="mw-page-title-main">Bibliometrics</span> Statistical analysis of written publications

Bibliometrics is the application of statistical methods to the study of bibliographic data, especially in scientific and library and information science contexts, and is closely associated with scientometrics to the point that both fields largely overlap.

Scientometrics is a subfield of informetrics that studies quantitative aspects of scholarly literature. Major research issues include the measurement of the impact of research papers and academic journals, the understanding of scientific citations, and the use of such measurements in policy and management contexts. In practice there is a significant overlap between scientometrics and other scientific fields such as information systems, information science, science of science policy, sociology of science, and metascience. Critics have argued that overreliance on scientometrics has created a system of perverse incentives, producing a publish or perish environment that leads to low-quality research.

Citation analysis is the examination of the frequency, patterns, and graphs of citations in documents. It uses the directed graph of citations — links from one document to another document — to reveal properties of the documents. A typical aim would be to identify the most important documents in a collection. A classic example is that of the citations between academic articles and books. For another example, judges of law support their judgements by referring back to judgements made in earlier cases. An additional example is provided by patents which contain prior art, citation of earlier patents relevant to the current claim. The digitization of patent data and increasing computing power have led to a community of practice that uses these citation data to measure innovation attributes, trace knowledge flows, and map innovation networks.

<span class="mw-page-title-main">Informetrics</span> Study of the quantitative aspects of information

Informetrics is the study of quantitative aspects of information, it is an extension and evolution of traditional bibliometrics and scientometrics. Informetrics uses bibliometrics and scientometrics methods to study mainly the problems of literature information management and evaluation of science and technology. Informetrics is an independent discipline that uses quantitative methods from mathematics and statistics to study the process, phenomena, and law of informetrics. Informetrics has gained more attention as it is a common scientific method for academic evaluation, research hotspots in discipline, and trend analysis.

Citation impact or citation rate is a measure of how many times an academic journal article or book or author is cited by other articles, books or authors. Citation counts are interpreted as measures of the impact or influence of academic work and have given rise to the field of bibliometrics or scientometrics, specializing in the study of patterns of academic impact through citation analysis. The importance of journals can be measured by the average citation rate, the ratio of number of citations to number articles published within a given time period and in a given index, such as the journal impact factor or the citescore. It is used by academic institutions in decisions about academic tenure, promotion and hiring, and hence also used by authors in deciding which journal to publish in. Citation-like measures are also used in other fields that do ranking, such as Google's PageRank algorithm, software metrics, college and university rankings, and business performance indicators.

The h-index is an author-level metric that measures both the productivity and citation impact of the publications, initially used for an individual scientist or scholar. The h-index correlates with success indicators such as winning the Nobel Prize, being accepted for research fellowships and holding positions at top universities. The index is based on the set of the scientist's most cited papers and the number of citations that they have received in other publications. The index has more recently been applied to the productivity and impact of a scholarly journal as well as a group of scientists, such as a department or university or country. The index was suggested in 2005 by Jorge E. Hirsch, a physicist at UC San Diego, as a tool for determining theoretical physicists' relative quality and is sometimes called the Hirsch index or Hirsch number.

Journal ranking is widely used in academic circles in the evaluation of an academic journal's impact and quality. Journal rankings are intended to reflect the place of a journal within its field, the relative difficulty of being published in that journal, and the prestige associated with it. They have been introduced as official research evaluation tools in several countries.

The Eigenfactor score, developed by Jevin West and Carl Bergstrom at the University of Washington, is a rating of the total importance of a scientific journal. Journals are rated according to the number of incoming citations, with citations from highly ranked journals weighted to make a larger contribution to the eigenfactor than those from poorly ranked journals. As a measure of importance, the Eigenfactor score scales with the total impact of a journal. All else equal, journals generating higher impact to the field have larger Eigenfactor scores. Citation metrics like eigenfactor or PageRank-based scores reduce the effect of self-referential groups.

A bibliometrician is a researcher or a specialist in bibliometrics. It is near-synonymous with an informetrican, a scientometrican and a webometrician, who study webometrics.

<span class="mw-page-title-main">Altmetrics</span> Alternative metrics for analyzing scholarship

In scholarly and scientific publishing, altmetrics are non-traditional bibliometrics proposed as an alternative or complement to more traditional citation impact metrics, such as impact factor and h-index. The term altmetrics was proposed in 2010, as a generalization of article level metrics, and has its roots in the #altmetrics hashtag. Although altmetrics are often thought of as metrics about articles, they can be applied to people, journals, books, data sets, presentations, videos, source code repositories, web pages, etc.

Johan Lambert Trudo Maria Bollen is a scientist investigating complex systems and networks, the relation between social media and a variety of socio-economic phenomena such as the financial markets, public health, and social well-being, as well as Science of Science with a focus on impact metrics derived from usage data. He presently works as Professor of Informatics and Cognitive Science at Indiana University Bloomington and a fellow at the SparcS Institute of Wageningen University and Research Centre in the Netherlands. He is best known for his work on computational social science, scholarly impact metrics, measuring public well-being from large-scale social media data, and correlating Twitter mood to stock market prices.

Article-level metrics are citation metrics which measure the usage and impact of individual scholarly articles. The most common article-level citation metric is the number of citations. Field-weighted Citation Impact (FWCI) by Scopus divides the total citations by the average number of citations for an article in the scientific field.

<span class="mw-page-title-main">Microsoft Academic</span> Online bibliographic database

Microsoft Academic was a free internet-based academic search engine for academic publications and literature, developed by Microsoft Research in 2016 as a successor of Microsoft Academic Search. Microsoft Academic was shut down in 2022. Both OpenAlex and The Lens claim to be successors to Microsoft Academic.

There are a number of approaches to ranking academic publishing groups and publishers. Rankings rely on subjective impressions by the scholarly community, on analyses of prize winners of scientific associations, discipline, a publisher's reputation, and its impact factor.

<span class="mw-page-title-main">Ronald Rousseau</span>

Ronald Rousseau is a Belgian mathematician and information scientist. He has obtained an international reputation for his research on indicators and citation analysis in the fields of bibliometrics and scientometrics.

The Leiden Manifesto for research metrics (LM) is a list of "ten principles to guide research evaluation", published as a comment in Volume 520, Issue 7548 of Nature, on 22 April 2015. It was formulated by public policy professor Diana Hicks, scientometrics professor Paul Wouters, and their colleagues at the 19th International Conference on Science and Technology Indicators, held between 3–5 September 2014 in Leiden, The Netherlands.

The science-wide author databases of standardized citation indicators is a multidimensional ranking of the world's scientists produced since 2015 by a team of researchers led by John P. A. Ioannidis at Stanford.

References

  1. 1 2 3 4 Hirsch, J. E. (7 November 2005). "An index to quantify an individual's scientific research output". Proceedings of the National Academy of Sciences. 102 (46): 16569–16572. arXiv: physics/0508025 . Bibcode:2005PNAS..10216569H. doi: 10.1073/pnas.0507655102 . PMC   1283832 . PMID   16275915.
  2. 1 2 Peter A., Lawrence (2007). "The mismeasurement of science" (PDF). Current Biology. 17 (15): R583–R585. Bibcode:2007CBio...17.R583L. doi:10.1016/j.cub.2007.06.014. PMID   17686424. S2CID   30518724.
  3. 1 2 Şengör, Celâl. AM (2014). "How scientometry is killing science" (PDF). GSA Today. 24 (12): 44–45. doi:10.1130/GSATG226GW.1.
  4. West, Jevin D.; Jensen, Michael C.; Dandrea, Ralph J.; Gordon, Gregory J.; Bergstrom, Carl T. (2013). "Author-level Eigenfactor metrics: Evaluating the influence of authors, institutions, and countries within the social science research network community". Journal of the American Society for Information Science and Technology . 64 (4): 787–801. doi:10.1002/asi.22790.
  5. 1 2 Pan, Raj Kumar; Fortunato, Santo (2014). "Author Impact Factor: Tracking the dynamics of individual scientific impact". Scientific Reports . 4: 4880. arXiv: 1312.2650 . Bibcode:2014NatSR...4E4880P. doi:10.1038/srep04880. PMC   4017244 . PMID   24814674.
  6. Batista P. D.; et al. (2006). "Is it possible to compare researchers with different scientific interests?". Scientometrics . 68 (1): 179–89. arXiv: physics/0509048 . doi:10.1007/s11192-006-0090-4. S2CID   119068816.
  7. Koltun, V; Hafner, D (2021). "The h-index is no longer an effective correlate of scientific reputation". PLOS ONE. 16 (6): e0253397. arXiv: 2102.03234 . Bibcode:2021PLoSO..1653397K. doi: 10.1371/journal.pone.0253397 . PMC   8238192 . PMID   34181681.
  8. Bornmann, Lutz; Mutz, Rüdiger; Daniel, Hans-Dieter (2010). "The h index research output measurement: Two approaches to enhance its accuracy". Journal of Informetrics. 4 (3): 407–14. doi:10.1016/j.joi.2010.03.005.
  9. Cooke, Bec. "Guides: Research Metrics: Field-Weighted Citation Impact". libguides.usc.edu.au.
  10. "Snowball Metrics Recipe Book" (PDF). 2012.
  11. Tauro, Kiera. "Subject Guides: 6. Measure Impact: Field-Weighted Citation Impact". canterbury.libguides.com.
  12. Kaur, Jasleen; Radicchi, Filippo; Menczer, Filippo (2013). "Universality of scholarly impact metrics". Journal of Informetrics. 7 (4): 924–32. arXiv: 1305.6339 . doi:10.1016/j.joi.2013.09.002. S2CID   7415777.
  13. West, Jevin D.; Jensen, Michael C.; Dandrea, Ralph J.; Gordon, Gregory J.; Bergstrom, Carl T. (April 2013). "Author-level Eigenfactor metrics: Evaluating the influence of authors, institutions, and countries within the social science research network community". Journal of the American Society for Information Science and Technology. 64 (4): 787–801. doi:10.1002/asi.22790.
  14. Kashyap Dixit; S Kameshwaran; Sameep Mehta; Vinayaka Pandit; N Viswanadham (February 2009). "Towards simultaneously exploiting structure and outcomes in interaction networks for node ranking" (PDF). IBM Research Report R109002.; see also Kameshwaran, Sampath; Pandit, Vinayaka; Mehta, Sameep; Viswanadham, Nukala; Dixit, Kashyap (2010). "Outcome aware ranking in interaction networks". Proceedings of the 19th ACM international conference on Information and knowledge management – CIKM '10. p. 229. doi:10.1145/1871437.1871470. ISBN   9781450300995.
  15. Connor, James; Google Scholar Blog. "Google Scholar Citations Open To All", Google, 16 November 2011, retrieved 24 November 2011
  16. ""Professoren der nächsten Generation" | NZZ". Neue Zürcher Zeitung (in German). Retrieved 25 May 2020.
  17. Knowles, Jamillah (10 August 2012). "ResearchGate Releases RG Score - Klout for Boffins". The Next Web. Retrieved 26 May 2020.
  18. 1 2 Thelwall, M.; Kousha, K. (2014). "ResearchGate: Disseminating, communicating, and measuring Scholarship?" (PDF). Journal of the Association for Information Science and Technology. 66 (5): 876–889. CiteSeerX   10.1.1.589.5396 . doi:10.1002/asi.23236. S2CID   8974197. Archived (PDF) from the original on 2018-02-18. Retrieved 2018-07-30.
  19. Yu, Min-Chun (February 2016). "ResearchGate: An effective altmetric indicator for active researchers?". Computers in Human Behavior. 55: 1001–1006. doi:10.1016/j.chb.2015.11.007.
  20. 1 2 Kraker, Peter; Lex, Elisabeth (2015). "A Critical Look at the ResearchGate Score as a Measure of Scientific Reputation". Proceedings of the Quantifying and Analysing Scholarly Communication on the Web Workshop (ASCW'15).
  21. 1 2 Jordan, Katy (2015). Exploring the ResearchGate score as an academic metric: Reflections and implications for practice. Quantifying and Analysing Scholarly Communication on the Web (ASCW'15).
  22. Hoffmann, C. P.; Lutz, C.; Meckel, M. (2016). "A relational altmetric? Network centrality on ResearchGate as an indicator of scientific impact" (PDF). Journal of the Association for Information Science and Technology. 67 (4): 765–775. doi:10.1002/asi.23423. S2CID   7769870.
  23. Yu, Min-Chun (February 2016). "ResearchGate: An effective altmetric indicator for active researchers?". Computers in Human Behavior. 55: 1001–1006. doi:10.1016/j.chb.2015.11.007.
  24. Anne-Wil Harzing (2008-04-23). "Reflections on the h-index" . Retrieved 2013-07-18.
  25. von Bohlen und Halbach O (2011). "How to judge a book by its cover? How useful are bibliometric indices for the evaluation of "scientific quality" or "scientific productivity"?". Annals of Anatomy . 193 (3): 191–96. doi:10.1016/j.aanat.2011.03.011. PMID   21507617.
  26. Egghe, Leo (2006). "Theory and practise of the g-index". Scientometrics. 69 (1): 131–152. doi:10.1007/s11192-006-0144-7. hdl: 1942/981 . S2CID   207236267.
  27. Zhang, Chun-Ting (2009). Joly, Etienne (ed.). "The e-Index, Complementing the h-Index for Excess Citations". PLOS ONE. 4 (5): e5429. Bibcode:2009PLoSO...4.5429Z. doi: 10.1371/journal.pone.0005429 . PMC   2673580 . PMID   19415119.
  28. Dodson, M.V. (2009). "Citation analysis: Maintenance of h-index and use of e-index". Biochemical and Biophysical Research Communications. 387 (4): 625–26. doi:10.1016/j.bbrc.2009.07.091. PMID   19632203.
  29. Bras-Amorós, M.; Domingo-Ferrer, J.; Torra, V (2011). "A bibliometric index based on the collaboration distance between cited and citing authors". Journal of Informetrics. 5 (2): 248–64. doi:10.1016/j.joi.2010.11.001. hdl:10261/138172.
  30. Dorogovtsev, S.N.; Mendes, J.F.F. (2015). "Ranking Scientists". Nature Physics . 11 (11): 882–84. arXiv: 1511.01545 . Bibcode:2015NatPh..11..882D. doi:10.1038/nphys3533. S2CID   12533449.
  31. Fatchur Rochim, Adian (November 2018). "Improving fairness of h-index: RA-index". DESIDOC Journal of Library & Information Technology. 38 (6): 378–386. doi: 10.14429/djlit.38.6.12937 .
  32. Belikov, Aleksey V.; Belikov, Vitaly V. (22 September 2015). "A citation-based, author- and age-normalized, logarithmic index for evaluation of individual researchers independently of publication counts". F1000Research. 4: 884. doi: 10.12688/f1000research.7070.1 . PMC   4654436 .
  33. Engine, exaly Search. "exaly Search Engine". Free database of papers and journals. Retrieved 20 May 2022.
  34. Silagadze, Z. K. (2010). "Citation entropy and research impact estimation". Acta Phys. Pol. B. 41 (2010): 2325–33. arXiv: 0905.1039 . Bibcode:2009arXiv0905.1039S.
  35. Wu, Qiang (2009). "The w-index: A measure to assess scientific impact by focusing on widely cited papers". Journal of the American Society for Information Science and Technology. 61 (3): 609–614. arXiv: 0805.4650 . doi:10.1002/asi.21276. S2CID   119293304.
  36. Tscharntke, T.; Hochberg, M. E.; Rand, T. A.; Resh, V. H.; Krauss, J. (2007). "Author Sequence and Credit for Contributions in Multiauthored Publications". PLOS Biology. 5 (1): e18. doi: 10.1371/journal.pbio.0050018 . PMC   1769438 . PMID   17227141.
  37. Gągolewski, M.; Grzegorzewski, P. (2009). "A geometric approach to the construction of scientific impact indices". Scientometrics. 81 (3): 617–34. doi:10.1007/s11192-008-2253-y. S2CID   466433.
  38. Acuna, Daniel E.; Allesina, Stefano; Kording, Konrad P. (2012). "Future impact: Predicting scientific success". Nature. 489 (7415): 201–02. Bibcode:2012Natur.489..201A. doi:10.1038/489201a. PMC   3770471 . PMID   22972278.
  39. Penner, Orion; Pan, Raj K.; Petersen, Alexander M.; Kaski, Kimmo; Fortunato, Santo (2013). "On the Predictability of Future Impact in Science". Scientific Reports. 3 (3052): 3052. arXiv: 1306.0114 . Bibcode:2013NatSR...3E3052P. doi:10.1038/srep03052. PMC   3810665 . PMID   24165898.
  40. Schreiber, Michael (2015). "Restricting the h-index to a publication and citation time window: A case study of a timed Hirsch index". Journal of Informetrics. 9: 150–55. arXiv: 1412.5050 . doi:10.1016/j.joi.2014.12.005. S2CID   12320545.
  41. Hirsch, Jorge E. (2020). "Superconductivity, What the H? The Emperor Has No Clothes". Physics and Society. 49: 5–9. arXiv: 2001.09496 . I proposed the H-index hoping it would be an objective measure of scientific achievement. By and large, I think this is believed to be the case. But I have now come to believe that it can also fail spectacularly and have severe unintended negative consequences. I can understand how the sorcerer's apprentice must have felt. (p.5)
  42. Seppelt, Ralf (2018). "The Art of Scientific Performance". Trends in Ecology and Evolution. 11 (33): 805–809. doi:10.1016/j.tree.2018.08.003. PMID   30270172. S2CID   52890068.
  43. Gálvez RH (March 2017). "Assessing author self-citation as a mechanism of relevant knowledge diffusion". Scientometrics. 111 (3): 1801–1812. doi:10.1007/s11192-017-2330-1. S2CID   6863843.
  44. Christoph Bartneck & Servaas Kokkelmans; Kokkelmans (2011). "Detecting h-index manipulation through self-citation analysis". Scientometrics . 87 (1): 85–98. doi:10.1007/s11192-010-0306-5. PMC   3043246 . PMID   21472020.
  45. Emilio Ferrara & Alfonso Romero; Romero (2013). "Scientific impact evaluation and the effect of self-citations: Mitigating the bias by discounting the h-index". Journal of the American Society for Information Science and Technology . 64 (11): 2332–39. arXiv: 1202.3119 . doi:10.1002/asi.22976. S2CID   12693511.
  46. Labbé, Cyril (2010). Ike Antkare one of the great stars in the scientific firmament (PDF). Laboratoire d'Informatique de Grenoble RR-LIG-2008 (technical report) (Report). Joseph Fourier University.
  47. Wilhite, A. W.; Fong, E. A. (2012). "Coercive Citation in Academic Publishing". Science. 335 (6068): 542–3. Bibcode:2012Sci...335..542W. doi:10.1126/science.1212540. PMID   22301307. S2CID   30073305.
  48. Noorden, Richard Van (February 6, 2020). "Highly cited researcher banned from journal board for citation abuse". Nature. 578 (7794): 200–201. Bibcode:2020Natur.578..200V. doi: 10.1038/d41586-020-00335-7 . PMID   32047304.
  49. Rustam Tagiew; Dmitry I. Ignatov (2017). "Behavior mining in h-index ranking game" (PDF). CEUR Workshop Proceedings. 1968: 52–61.
  50. Fire, Michael; Guestrin, Carlos (1 June 2019). "Over-optimization of academic publishing metrics: observing Goodhart's Law in action". GigaScience. 8 (6). arXiv: 1809.07841 . doi:10.1093/gigascience/giz053. PMC   6541803 . PMID   31144712.
  51. 1 2 The Voice of the Dolphins and Other Stories. New York: Simon and Schuster. 1961.
  52. Committee, United States Congress Senate Appropriations (1961). Labor-Health, Education, and Welfare Appropriations for 1962, Hearings Before the Subcommittee of ... , 87-1 on H.R. 7035. p. 1498.

Further reading