Convergent validity of several indicators measuring disruptiveness with milestone assignments to physics papers by experts

https://doi.org/10.1016/j.joi.2021.101159Get rights and content

Highlights

  • Disruptive research diverges from current lines of research by opening up new lines.

  • We investigated the convergent validity of several indicators measuring disruptiveness.

  • We used a list of milestone papers, selected and published by editors of Physical Review Letters.

  • We investigated whether this editor-based list is related to values of the disruption indicators variants.

  • The results show that the indicators correlate differently with the milestone paper assignments by the editors.

Abstract

This study focuses on a recently introduced type of indicator measuring disruptiveness in science. Disruptive research diverges from current lines of research by opening up new lines. In the current study, we included the initially proposed indicator of this new type (Funk & Owen-Smith, 2017; Wu, Wang, & Evans, 2019) and several variants with DI1: DI5, DI1n, DI5n, and DEP. Since indicators should measure what they propose to measure, we investigated the convergent validity of the indicators. We used a list of milestone papers, selected and published by editors of Physical Review Letters, and investigated whether this human (experts)-based list is related to values of the several disruption indicators variants and – if so – which variants show the highest correlation with expert judgements. We used bivariate statistics, multiple regression models, and (coarsened) exact matching (CEM) to investigate the convergent validity of the indicators. The results show that the indicators correlate differently with the milestone paper assignments by the editors. It is not the initially proposed disruption index that performed best (DI1), but the variant DI5 which has been introduced by Bornmann, Devarakonda, Tekles, and Chacko (2020a). In the CEM analysis of this study, the DEP variant – introduced by Bu, Waltman, and Huang (in press) – also showed favorable results.

Introduction

The quantitative study of science using bibliometric indicators (publication and citation data) is widespread today (Fortunato et al., 2018). Research into science is aimed “at advancing our knowledge on the development of science and its communication structure, as well as in relation to social, technological, and socioeconomic aspects” (van Raan, 2019, p. 238). This research mainly began in the 1960s when Eugene Garfield created the Science Citation Index and developed into the broad activity of many scientists with the launch of the Web of Science (WoS, now Clarivate Analytics; Wilsdon et al., 2015). Professional bibliometrics is characterized by the use of advanced bibliometric indicators (Leydesdorff, Wouters, & Bornmann, 2016; van Raan, 2019). These indicators mainly refer to field-normalized indicators such as citation percentiles (Bornmann, 2019). However, the use of these indicators in research evaluation resulted in the critique that “citation-based funding could favor mainstream research and bias decisions against more original projects” (Jappe, Pithan, & Heinze, 2018). Citation counts (and indicators derived from these) might not be able to identify discoveries that are defined by Ziman (1987) as follows: “research results that make a significant change in what we thought we already knew” (p. 98). Highly cited papers might follow short-lasting trends, review the existing literature, or introduce research methods and algorithms (van Raan, 2019).

In recent years, several alternative indicators to citations have been developed measuring novelty of research (Bornmann, Tekles, Zhang, & Ye, 2019). Novelty is related to creativity (novelty is an ingredient of creativity, see Puccio, Mance, & Zacko-Smith, 2013) that involves “the production of high-quality, original, and elegant solutions to complex, novel, ill-defined, or poorly structured problems” (Hemlin, Allwood, Martin, & Mumford, 2013, p. 10). The newly developed novelty indicators are based on the view of creativity as novel recombination of existing elements. For example, Uzzi, Mukherjee, Stringer, and Jones (2013) interpreted more frequent unusual cited references (cited journal) combinations as hints to more novel knowledge (see Lee, Walsh, & Wang, 2015). An overview of the different available approaches in scientometrics to measuring novelty based on unusual combinations of existing elements (e.g. cited references or key words) can be found in Wang, Lee, and Walsh (2018) and Wagner, Whetsell, and Mukherjee (2019).

Based on the fact that impact metrics only measure possible use of new ideas, Funk and Owen-Smith (2017) introduced an indicator (based on patent citations) that is intended to measure newness challenging the existing order. Wu, Wang, & Evans (2019) transferred the idea by Funk and Owen-Smith (2017) to bibliometrics by introducing a disruption index (Shibayama & Wang, 2019, proposed similar measures). Azoulay (2019) explains the intuition behind the new indicator as follows: “when the papers that cite a given article also reference a substantial proportion of that article’s references, then the article can be seen as consolidating its scientific domain. When the converse is true – that is, when future citations to the article do not also acknowledge the article’s own intellectual forebears – the article can be seen as disrupting its domain”. A similar index (the so-called dependency indicator, DEP) has been proposed by Bu, Waltman, and Huang (2021), who introduced the dependency indicator from a multi-dimensional perspective of impact measurement. According to this perspective, further information from the citing and cited side is considered to measure performance (which is in contrast to the one-dimensional times cited indicator).

The various disruption indicator variants are connected to the distinction by Kuhn (1962) between normal science and scientific revolutions. Kuhn (1962) presents a theory of scientific knowledge: according to Wray (2017), in this theory, “growth of science is not a continuous march closer and closer to the truth. Instead, periods of rapid growth are interrupted by revolutionary changes of theory. These revolutionary changes of theory are disruptive” (p. 66; see also Casadevall & Fang, 2016). In the context of Kuhn’s theory of scientific knowledge, Foster, Rzhetsky, and Evans (2015) differentiate between two strategies of conducting research (see also Merton, 1957). The different strategy poles can be denoted as traditional and risky innovative thinking or as convergent and divergent thinking. Divergent thinking increases the chance of conducting breakthrough research. Various papers have been published in scientometrics to date that focus on identifying breakthrough, landmark, or milestone papers (e.g. Chen, 2004; Schneider & Costas, 2016; Thor, Bornmann, Haunschild, & Leydesdorff, 2021). An overview of these papers can be found in Winnink, Tijssen, and van Raan (2016).

In this paper, we follow the approach of Bornmann, Devarakonda, Tekles, and Chacko (2020) and investigate the convergent validity of the new disruption indicators. We are interested in the question of whether the indicators measure what they intend to measure, namely the disruptiveness of research. The journal Physical Review Letters (PRL) published a list of milestone papers from the journal. The list can be used to validate the indicators: one can expect that milestone papers have higher indicator values than PRL papers not selected for the list. We investigated whether the indicators are able to identify milestone papers and whether there are differences observable between the indicators in this ability.

Section snippets

Dataset used

We retrieved the list of milestone papers published in PRL from https://journals.aps.org/prl/50years/milestones. According to a personal communication with Reinhardt B. Schuhmann, the current managing editor of PRL, and Anonymous (2008), there was an effort to cover all areas of physics research in the list. Many papers have been denoted as milestone papers since they are closely connected to Nobel Prizes in Physics, and occasionally in Chemistry. It cannot be ruled out that bibliometric

Results

In this study, we included four variants of the disruption index and citation counts. We are interested in the question as to how these performance measures correspond to the qualitative milestone paper assessments by the PRL. Fig. 2 shows two distributions of the disruptive index variants and citations: (1) the bars are actual distributions, and (2) the smooth, bell-shaped distribution outlines how the data would be distributed if they were normal. We transformed two variables in Fig. 2 for

Discussion

In the ethos of science, Merton (1973) included a set of norms that would guide the behavior of scientists. According to Ziman (2000), “the norms of ‘originality’, ‘scepticism’ and ‘communalism’ are put into operation as processes of ‘variation’, ‘selection’ and ‘retention’ respectively” (p. 277). The norm of originality ‘reminds’ researchers that it is their superficial role to advance knowledge: “in the institution of science originality is at premium. For it is through originality, in

Author contribution

Lutz Bornmann: Conceptualization, Methodology, Writing - original draft preparation, Writing - reviewing and editing, Data curation, Formal analysis, Validation. Alexander Tekles: Conceptualization, Methodology, Writing - original draft preparation, Writing - reviewing and editing, Data curation, Formal analysis, Validation.

CRediT authorship contribution statement

Lutz Bornmann: Conceptualization, Methodology, Writing - original draft, Writing - review & editing, Data curation, Formal analysis, Validation. Alexander Tekles: Conceptualization, Methodology, Writing - original draft, Writing - review & editing, Data curation, Formal analysis, Validation.

Acknowledgements

The bibliometric data used in this paper are from an in-house database developed and maintained in cooperation with the Max Planck Digital Library (MPDL, Munich) and derived from the Science Citation Index Expanded (SCI-E), Social Sciences Citation Index (SSCI), Arts and Humanities Citation Index (AHCI) prepared by Clarivate Analytics (Philadelphia, Pennsylvania, USA). We thank Henry Small for providing feedback on earlier versions of this paper.

References (94)

  • M.A. Schilling et al.

    Recombinant search and breakthrough idea generation: An analysis of high impact papers in the social sciences

    Research Policy

    (2011)
  • I. Tahamtan et al.

    Core elements in the process of citing publications: Conceptual overview of the literature

    Journal of Informetrics

    (2018)
  • I. Tahamtan et al.

    Creativity in science and the link to cited references: Is the creative potential of papers reflected in their cited references?

    Journal of Informetrics

    (2018)
  • T. Tregenza

    Gender bias in the refereeing process?

    Trends in Ecology & Evolution

    (2002)
  • C.S. Wagner et al.

    International research collaboration: Novelty, conventionality, and atypicality in knowledge recombination

    Research Policy

    (2019)
  • J. Wang et al.

    Funding model and creativity in science: Competitive versus block funding and status contingency effects

    Research Policy

    (2018)
  • A.C. Acock

    A gentle introduction to Stata

    (2018)
  • J. Adams et al.

    Interdisciplinary research: Methodologies for identification and assessment. Do we know what we are measuring?

    (2016)
  • P. Ahlgren et al.

    Exploring the relation between referencing practices and citation impact: A large‐scale study based on Web of Science data

    Journal of the Association for Information Science and Technology

    (2018)
  • D.W. Aksnes et al.

    Citations, citation indicators, and research quality: An overview of basic concepts and theories

    SAGE Open

    (2019)
  • G. Angeles et al.

    Guide to longitudinal program impact evaluation

    (2014)
  • Anon

    PRL showcases milestone papers

    APS News

    (2008)
  • P.C. Austin

    An introduction to propensity score methods for reducing the effects of confounding in observational studies

    Multivariate Behavioral Research

    (2011)
  • P. Azoulay

    Small research teams ‘disrupt’ science more radically than large ones

    Nature

    (2019)
  • A. Baccini et al.

    Do they agree? Bibliometric evaluation versus informed peer review in the Italian research assessment exercise

    Scientometrics

    (2016)
  • D.B. Beaver

    Does collaborative research have greater epistemic authority?

    Scientometrics

    (2004)
  • L. Bornmann

    Bibliometric indicators

  • L. Bornmann et al.

    Reviewer and editor biases in journal peer review: An investigation of manuscript refereeing at Angewandte Chemie International Edition

    Research Evaluation

    (2009)
  • L. Bornmann et al.

    Disruption index depends on length of citation window

    El profesional de la información

    (2019)
  • L. Bornmann et al.

    Are disruption index indicators convergently valid? The comparison of several indicator variants with assessments by peers

    Quantitative Science Studies

    (2020)
  • L. Bornmann et al.

    Disruptive papers published in Scientometrics: Meaningful results by using an improved variant of the disruption index originally proposed by Wu, Wang, and Evans (2019)

    Scientometrics

    (2020)
  • T.S. Breusch et al.

    A simple test for heteroscedasticity and random coefficient variation

    Econometrica

    (1979)
  • Y. Bu et al.

    A multi-dimensional framework for characterizing the citation impact of scientific publications

    Quantitative Science Studies

    (2021)
  • M. Caliendo et al.

    Some practical guidance for the implementation of propensity score matching

    Journal of Economic Surveys

    (2008)
  • A. Casadevall et al.

    Revolutionary science

    mBio

    (2016)
  • C. Chen

    Searching for intellectual turning points: Progressive knowledge domain visualization

    Proceedings of the National Academy of Sciences of the United States of America

    (2004)
  • W. Chunli et al.

    Nobel-Prize-winning papers are significantly more highly-cited but not more disruptive than non-Prize-winning counterparts

  • A. Colin Cameron et al.

    A practitioner’s guide to cluster-robust inference

    The Journal of Human Resources

    (2015)
  • R.D. Cook et al.

    Diagnostics for heteroscedasticity in regression

    Biometrika

    (1983)
  • R.B. Dagostino et al.

    A suggestion for using powerful and informative tests of normality

    The American Statistician

    (1990)
  • P. Dattalo

    Determining sample size: Balancing power, precision, and practicality

    (2008)
  • F. Didegah et al.

    Determinants of research citation impact in nanoscience and nanotechnology

    Journal of the American Society for Information Science and Technology

    (2013)
  • M.E. Falagas et al.

    The impact of article length on the number of future citations: A bibliometric analysis of general medicine journals

    PloS One

    (2013)
  • F. Ferretti et al.

    Research excellence indicators: Time to reimagine the ‘making of’?

    Science & Public Policy

    (2018)
  • S. Fortunato et al.

    Science of science

    Science

    (2018)
  • J.G. Foster et al.

    Tradition and innovation in scientists’ research strategies

    American Sociological Review

    (2015)
  • R.J. Funk et al.

    A dynamic network measure of technological change

    Management Science

    (2017)
  • Cited by (22)

    View all citing articles on Scopus
    View full text