Performance Measurement and Metrics

Similar documents
Resource Review. In press 2018, the Journal of the Medical Library Association

Outlining an analytical framework for mapping research evaluation landscapes 1

Increased Visibility in the Social Sciences and the Humanities (SSH)

1. Is Your Article Relevant to the Journal?

Can we better support and motivate scientists to deliver impact? Looking at the role of research evaluation and metrics. Áine Regan & Maeve Henchion

U-Multirank 2017 bibliometrics: information sources, computations and performance indicators

Infrastructures as analytical framework for mapping research evaluation landscapes and practices

STI 2018 Conference Proceedings

New forms of scholarly communication Lunch e-research methods and case studies

Title: The Impact of Altmetrics on library for biomedical research

Research assessment and evaluation in Russian fundamental science

Title: Can we innovate how we measure scientific impact?

Leiden Manifesto impact? Presented at American Evaluation Association Conference Atlanta, GA October 28, 2016

STI 2018 Conference Proceedings

Academy of Social Sciences response to Plan S, and UKRI implementation

Altmetrics could enable scholarship from developing countries to receive due recognition.

Research Excellence Framework

PLOS. From Open Access to Open Science : a publisher s perspective. Véronique Kiermer Executive Editor, PLOS Public Library of Science.

GUIDELINES SOCIAL SCIENCES AND HUMANITIES RESEARCH MATTERS. ON HOW TO SUCCESSFULLY DESIGN, AND IMPLEMENT, MISSION-ORIENTED RESEARCH PROGRAMMES

Reduce cost sharing and fees Include other services. Services: which services are covered? Population: who is covered?

Mother Jacobs Home Remedies, Now with an Economics Flavour: Tracking Jane Jacobs Influence using Bibliometric and Network Analysis

Taylor & Francis journals Canadian researcher survey 2010

Publishing in academic journals. Tips to help you succeed

Competing for Excellence: Perverse and constructive uses of evaluation machines in academia

New perspectives on article-level metrics: developing ways to assess research uptake and impact online

Working Paper Series of the German Data Forum (RatSWD)

Next generation research evaluation:!!!!!!!!!!! the ACUMEN Portfolio and web based information tools

CONSIDERATIONS REGARDING THE TENURE AND PROMOTION OF CLASSICAL ARCHAEOLOGISTS EMPLOYED IN COLLEGES AND UNIVERSITIES

S E R B A N I O N E S C U M. D. P H. D. U N I V E R S I T É P A R I S 8 U N I V E R S I T É D U Q U É B E C À T R O I S - R I V I È R E S

Issues in Emerging Health Technologies Bulletin Process

esss Berlin, 8 13 September 2013 Monday, 9 October 2013

The changing role for European research libraries

Audit culture, the enterprise university and public engagement

Tracking and predicting growth of health information using scientometrics methods and Google Trends

Big data for the analysis of digital economy & society Beyond bibliometrics

SciELO SA: Past, Present and Future (September 2018)

Introducing Elsevier Research Intelligence

Sanna Talja & Pertti Vakkari Scholarly publishing orientations and patterns of print and electronic literature use

SUPPORTING THE JOURNAL SELECTION PROCESS & RESEARCH PUBLICATION PRACTICES FOR RESEARCH PERFORMANCE EVALUATION IN SERBIA TITLE

Panel 1: Teaching, Publishing, Advising Issues of Quality

Publishing in academic journals. Tips to help you succeed

Researchers and new tools But what about the librarian? mendeley.com

Editors and Journal Startup in the Digital Era

Evolution of the Development of Scientometrics

Energy for society: The value and need for interdisciplinary research

On Epistemic Effects: A Reply to Castellani, Pontecorvo and Valente Arie Rip, University of Twente

The impact of the Online Knowledge Library: Its Use and Impact on the Production of the Portuguese Academic and Scientific Community ( )

Publishing for Impact

Trust, but Verify : What the Digital and Transparency Revolutions in Social Science Mean for You. Andrew Moravcsik

Towards a Software Engineering Research Framework: Extending Design Science Research

Bibliometrics between ambition and responsibility

A new Journal in the field of Innovation Management Editorial Kick-off CERN 31 st May 2016

In this presentation, I will briefly:

The role of SciELO on the road towards the Professionalization, Internationalization and Financial Sustainability of developing country journals

Grand Challenges in Measuring and Characterizing Scholarly Impact

Guidelines for the Professional Evaluation of Digital Scholarship by Historians

The impact of the Online Knowledge Library: its use and impact on the production of the Portuguese academic and scientific community ( )

Elements of Scholarly Discourse in a Digital World

JOURNAL PUBLISHING IN ASTRONOMY

If Our Research is Relevant, Why is Nobody Listening?

2. Overall Use of Technology Survey Data Report

Program Level Learning Outcomes for the Department of International Studies Page 1

Science Integration Fellowship: California Ocean Science Trust & Humboldt State University

Bold communication, responsible influence. Science communication recommendations

Fashion Technology Research: A Scientometric Analysis

The modern global researcher:

special roundtable Andrew D. Marble Kenneth Lieberthal Emily O. Goldman Robert Sutter Ezra F. Vogel Celeste A. Wallander

INVOLVING USERS TO SUCCESSFULLY MEET THE CHALLENGES OF THE DIGITAL LIBRARY: A 30 YEAR PERSONAL REFLECTION

Scientific publications as boundary objects: theorising the intersection of classification and research evaluation

Assessing progress. Impact of Science. Pillar Hall. Alliances with stakeholders. 12 & 13 June 2017, Stockholm. Partners & Supporters:

Centre for the Study of Human Rights Master programme in Human Rights Practice, 80 credits (120 ECTS) (Erasmus Mundus)

Exploring alternative cyberbibliometrics for evaluation of scholarly performance in the social sciences and humanities in Taiwan

Strategic Plan Public engagement with research

and its repercussions on the scientific community Berlin, September 2013 Koenraad Debackere, KU Leuven

Information Systems Frontiers CALL FOR PAPERS. Special Issue on: Digital transformation for a sustainable society in the 21st century

How to Evidence and Record Policy Impact

Plum Goes Orange Elsevier Acquires Plum Analytics - The Scho...

Early insights of Emerging Sources Citation Index (ESCI): a bibliometrics analysis and overlap mapping method

A selective list of sociology journals suitable for qualitative paper submission

A COMPREHENSIVE DATABASE OF HIGH-QUALITY RESEARCH. natureindex.com. Track top papers Explore collaborations Compare research performance

1. Introduction and About Respondents Survey Data Report

A MODEL OF SCHOLARLY COMMUNICATION IN TOURISM AND AN OPEN ACCESS INITIATIVE

Global Trends in Neuroscience Publishing Background and Developments

Constants and Variables in 30 Years of Science and Technology Policy. Luke Georghiou University of Manchester Presentation for NISTEP 30 Symposium

STI 2018 Conference Proceedings

The Importance of Scientific Reproducibility in Evidence-based Rulemaking

Association of American Publishers AAP

Next-generation metrics: Responsible metrics and evaluation for open science

Supporting medical technology development with the analytic hierarchy process Hummel, Janna Marchien

Publishing open access: a guide for authors

Some Thoughts On Peer Review In The Global Internet Context

Investigating the Determinants and Perceived Consequences of Corporate Auditory

Data integration in Scandinavia

Office of Science and Technology Policy th Street Washington, DC 20502

Polona Vilar, Primož Južnič, and Tomaz Bartol ; University of Ljubljana, Slovenia

Scandinavian versus UK research: The importance of institutional context

Quality in legal science: the case of evaluating legal monographs

The compliance of Iranian library and information science journals with Thomson Reuters basic standards

Data Science Research Fellow

CHAPTER 8 RESEARCH METHODOLOGY AND DESIGN

Transcription:

Principles to guide reliable and ethical research evaluation using metric-based indicators of impact Journal: Manuscript ID PMM-0-0-00 Manuscript Type: Research Paper Keywords: altmetrics, impact, metrics, metric-based indicators, academic pratice, scholarly impact

Page of 0 0 0 0 0 0 Principles to guide reliable and ethical research evaluation using metric-based indicators Bibliometrics can be defined as a set of tools and techniques which enable quantitative analyses of scholarly literature. The analyses can be conducted for a variety of purposes from collection development and evaluation in research libraries, to tracking changes in scholarly disciplines, to studying social and organizational structures of science or assessing the role and/or contribution of individual researchers, research groups, institutions or nations. These studies were traditionally the domain of bibliometric researchers and scholars. They had access to often expensive data sets and worked to develop and validate bibliometric methods as well as to understand their limitations. This community created formal structures of discourse including specialized peerreviewed journals (e.g., Scientometrics, Research Evaluation, Journal of the Association for Information Science and Technology, and Journal of Informetrics), conferences (e.g., the Science and Technology Indicators Conference and the Conference of the International Society for Scientometrics and Informetrics), and societies (e.g., the International Society for Scientometrics and Informetrics). More recently, bibliometric tools have been employed to inform assessment of the quality and impact of research, either in an attempt to replace or to serve alongside the peer review process. We can consider citations as a form of peer review if we subscribe to the Mertonian theory of the normative structure of science. This theory proposes that scientific progress is possible only if scientists follow certain accepted norms of behavior, including organized skepticism []. Keen to understand more about the research they fund and conduct, research institutions, funders and government agencies turned to bibliometrics for a bias-free and inexpensive assessment

Page of 0 0 0 0 0 0 method. With the development of widely available on-line commercial tools, bibliometrics has become a method available to all, not just a select group of expert scholars. The journal impact factor (JIF) and h-index are familiar indicators to thousands of researchers around the world who need to demonstrate the impact of their past work. Promotion and Tenure Committees and organizations that fund research regularly ask applicants to provide the computed values of these indicators, and then use them in ways for which they were never intended. For example, the JIF tells us nothing about the quality of an individual paper in that journal. Anthony van Raan, former director of the Centre for Science and Technology Studies at Leiden University in the Netherlands, noted: If there is one thing every bibliometrician agrees, it is that you should never use the journal impact factor to evaluate research performance for an article or for an individual that is a mortal sin. [] Equally, comparing h-indices of researchers at different stages in their careers or across different disciplines can result in grossly misleading findings, which vary depending on the underlying data []. The research community affected by these practices is increasingly concerned about the inappropriate use of bibliometrics for evaluation and decision-making. Critiques have led, among others, to the Declaration on Research Assessment (DORA) declaration [], which called for the abandonment of the use of JIF for evaluation of individuals. Others have claimed that the current evaluation regimes in many countries, such as the United Kingdom [,,], Brazil [], and Australia [], may lead to distortions in scholars behaviors, leading to decreased creativity, risk aversion and less willingness to undertake interdisciplinary research.

Page of 0 0 0 0 0 0 Against this backdrop, research libraries are taking an increasingly proactive role in either supporting organization-wide evaluation efforts or developing bibliometric services to help researchers navigate the maze of research metrics and use them effectively. For instance, in Australia and the United Kingdom, libraries are funded to collect and verify research outputs submitted by institutions for national assessment. In the United States and elsewhere, libraries are launching new services to help researchers and their organizations understand the impact of their research. The Association of Research Libraries (ARL) SPEC survey [0] reported that in January 0 there were ARL-member libraries with such services either already developed or being developed. Developments continue, such as at the University of Waterloo in Canada, which recently released a guide on bibliometric measures, together with a description of available tools []. With very few exceptions, these services are not run by bibliometrics researchers. A panel entitled How to Deal with Unsettling Realities of Bibliometric Services in Universities addressed this very issue at the 0 Science and Technology Indicators Conference in Leiden []. THE LEIDEN MANIFESTO At that same conference, bibliometric researchers and practitioners from around the globe issued the so-called Leiden Manifesto []. This document, published in Nature, laid down ten principles intended to guide best practices for bibliometric-based research assessment. These principles are summarized below. Although they are well understood within the expert bibliometric community, they need to be adopted by others seeking to implement or understand bibliometric-based evaluation. The Association of Research Libraries (ARL) is a group of research libraries across the United States and Canada.

Page of 0 0 0 0 0 0 Principle. Metrics can provide additional dimensions to the assessment process, but should never be used in isolation from qualitative assessment (e.g., peer-review). Metrics-based evaluation can supplement and provide additional dimensions to qualitative assessment, but should never replace it. Principle. Metrics used to evaluate research performance should reflect the research objectives of the institution, research groups or individual researchers. Individual indicators often provide a one-dimensional view of research impact while intended research goals of the evaluated units or individuals may be multi-dimensional. For example, they may include advances of science or improvements of social outcomes and may be aimed at differing audiences from researcher, to industry, to policy makers. No single metric or evaluation model can apply in all contexts. Principle. Measure locally relevant research using appropriate metrics, including those that build on journal collections in local languages or that cover certain geographic locations. Big international citation databases (used most frequently to derive data used for constructing indicators) still mostly focus on English-language, western journals. Principle. Metrics-based evaluation, to be trusted, should adhere to the standards of openness and transparency in data collection and analysis. What data are collected? How is it collected? How are citations captured? What are the exact methods and calculations used to develop indicators? Is the process open to scrutiny by experts and by the assessed? Principle. Those who are evaluated should be able to verify data and the analyses used in the assessment process. Are all relevant outputs identified, captured and analyzed?

Page of 0 0 0 0 0 0 Principle. Just as all metrics are not suitable for assessing all aspects of scholarship (see Principle ) neither can they be applied equally across all disciplines. We know that disciplines vary in their publication and citation practices, and these need to be taken into consideration when selecting metrics to compare disciplines. For instance, a bibliometric profile of a researcher studying causes of lung disease will be rather different from that of a researcher studying the social effects of smoke cessation programs. Health policy research tends to behave more in line with the softer disciplines, with fewer citation counts, a more diffuse set of outlets, and top journals with lower impact factor values. In contrast, biomedical research tends to behave more in line with the harder sciences (i.e., with higher average citation rates). For instance, the topranking health policy journal has a JIF of. while the top respiratory system journal has a JIF of.. If comparisons across disciplines are called for, the most suitable metrics are those that statistically normalize for disciplinary differences. They should compare to the discipline baselines. For instance, a researcher might ask, "Do my hematology publications have more citations than an average hematology publication of the same age? In a percentile distribution of all hematology publications, based on citation counts, are my publications in the th, th or 0 th percentile?" Principle. Do not rely on a single quantitative indicator when evaluating individual researchers. The h-index, currently the most popular author-level indicator, favors older researchers with longer publication lists. Moreover, it does not adjust for disciplinary differences and ignores the The 0 Journal Citation Report shows Health Affairs with a JIF of. and American Journal of Respiratory and Critical Care Medicine a JIF of..

Page of 0 0 0 0 0 0 impact of highly cited papers. The signatories of the Leiden Manifesto state that: Reading and judging a researcher's work is much more appropriate than relying on one number. [] Principle. Sets of indicators can provide a more reliable and multi-dimensional view than a single indicator. The Manifesto authors give an example of a set of impact factors shown to three decimal places, creating a false impression that journals can be reliably ranked even if small differences in scores are observed. It is better to consider a range of indicators to identify differences. Principles and 0. Goodhart s Law is evident in research evaluation; it states that, any observed statistical regularity will tend to collapse once pressure is placed upon it for control purposes. [] Every evaluation system creates incentives (intended or unintended) and these, in turn, drive behaviors. Use of a single indicator (like JIF) opens the evaluation system to such undesirable behaviors like gaming or goal displacement. To mitigate against these behaviors multiple indicators should be used. Furthermore, indicators should be reviewed and updated in line with changing goals of assessment, and new metrics should be considered as they become available. While I believe that libraries in research institutions are well placed to provide institutional support for metric-based evaluation, I also strongly believe that librarians should become advocates for the responsible and ethical use of these metrics. A more popular version of the Goodhart s Law reads when a measure becomes a target, it ceases to be a good measure. A good example of gaming are so-called citation clubs designed to artificially increase citation rates to publications, while goal displacement refers the behaviors in which the measurement becomes a goal (e.g., the only criterion for the selection of a publication outlet is its impact factor).

Page of 0 0 0 0 0 0 REFERENCES. Merton, R.K. The Sociology of Science: Theoretical and Empirical Investigations. Chicago: University of Chicago Press,. Chapter, The Normative Structure of Science, p. -.. Van Noorden R. Metrics: A profusion of measures. Nature 00 June ; (00):. doi:0.0/a. Bar Ilan, J. Which h-index? A comparison of WoS, Scopus and Google Scholar. Scientometrics 00 ():-. doi:0.00/s-00-0-y. San Francisco Declaration on Research Assessment (DORA) http://www.ascb.org/dora/. de Rijcke, S., Woulters, P., Rushforth, A.D., Franssen, T.P., Hammarfelt, B. Evaluation practices and effects of indicator use a literature review. Res Eval. 0. Advanced access Dec. 0 doi:0.0/reseval/rvv0. Moriarty, P. Addicted to the brand: the hypocrisy of a publishing academic. LSE Impact Blog. 0 March. Available at: http://blogs.lse.ac.uk/impactofsocialsciences/0/0//addicted-to-the-brand-thehypocrisy-of-a-publishing-academic/. Shaw. C. Research that does not belong to a single subject area deemed too risky. The Guardian (US edition). 0 Nov Available at: http://www.theguardian.com/highereducation-network/blog/0/nov//interdisciplinary-research-ref-submission-university

Page of 0 0 0 0 0 0. Ferreira R.C., Antoneli, F., Briones, M.R.S. The hidden factors in impact factors: A perspective from Brazilian science. Front Genet. 0 July ; (art.0). doi:0./fgene.0.000. Kwok, J.T. Impact of ERA Research Assessment on University Behaviour and their Staff. Melbourne: NTEU National Policy and Research Unit, 0. Available at: http://apo.org.au/files/resource/nteu_impactofera_april0.pdf 0. SPEC Kit : Scholarly Output Assessment Activities. May 0. Available at: http://publications.arl.org/scholarly-output-assessment-spec-kit-/. University of Waterloo Working Group on Bibliometrics. White Paper on Bibliometrics, Measuring Research Productivity and Impact Through Bibliometrics. 0 Waterloo, Ontario: University of Waterloo. Available at: https://uwaterloo.ca/institutional-analysis-planning/sites/ca.institutionalanalysisplanning/files/uploads/files/white_paper_on_bibliometrics_draft_for_consultatio n_0oct0_0.pdf. Drenthe, G. How to Deal with Unsettling Realities of Bibliometric Services in Universities. Special Panel during 0 STI Conference, Leiden.. Hicks, D., Wouters, P., Waltman, L., de Rijcke, S., Rafols, I. The Leiden Manifesto for Research Metrics. Nature 0 April ; 0():-. Available at: http://www.nature.com/polopoly_fs/.!/menu/main/topcolumns/topleftcolumn/pd f/0a.pdf. Goodhart, C.A.E. Monetary Theory and Practice: The UK experience. London: The Macmillan Press,. Chapter Problems of Monetary Management: The UK Experience, p. -.

Page of 0 0 0 0 0 0