Research assessment and evaluation in Russian fundamental science

Similar documents
Constants and Variables in 30 Years of Science and Technology Policy. Luke Georghiou University of Manchester Presentation for NISTEP 30 Symposium

The changing role for European research libraries

Benchmarking National Innovation Capability: Indicators Framework and Primary Findings

Research Excellence Framework

SERBIA. National Development Plan. November

Leiden Manifesto impact? Presented at American Evaluation Association Conference Atlanta, GA October 28, 2016

Performance Measurement and Metrics

Innovative Models of Teaching in Training of Adolescents Chess Players

SR&ED International R&D Tax Credit Strategies

ENUMERATE: Measuring the progress of digital heritage in Europe

COUNTRY REPORT: TURKEY

U-Multirank 2017 bibliometrics: information sources, computations and performance indicators

SUPPORTING THE JOURNAL SELECTION PROCESS & RESEARCH PUBLICATION PRACTICES FOR RESEARCH PERFORMANCE EVALUATION IN SERBIA TITLE

PRINCIPLES AND CRITERIA FOR THE EVALUATION OF SCIENTIFIC ORGANISATIONS IN THE REPUBLIC OF CROATIA

AC : THE NATIONAL PROJECT FOR THE INNOVATIVE DEVELOPMENT OF THE HIGHER ENGINEERING EDUCATION IN RUSSIA

Data integration in Scandinavia

Agricultural Data Verification Protocol for the Chesapeake Bay Program Partnership

Appendix B: Example Research-Activity Description

Capturing and Conveying the Essence of the Space Economy

How CRISs are key to the future of research libraries INCONECSS April 2016 Berlin

Latin-American non-state actor dialogue on Article 6 of the Paris Agreement

Canada Agricultural Census 2011 Explanatory notes

CONFERENCE AND JOURNAL TRANSPORT PROBLEMS. WHAT'S NEW?

Increased Visibility in the Social Sciences and the Humanities (SSH)

ANNEXES FOLLOW-UP OF RECOMMENDATIONS BY ORDER OF PRIORITY

Committee on Development and Intellectual Property (CDIP)

Big data for the analysis of digital economy & society Beyond bibliometrics

Introduction to HSE ISSEK

Research & Development (R&D) defined (3 phase process)

Introducing Elsevier Research Intelligence

Russian Scientific Heritage Digital Library

Provided by. RESEARCH ON INTERNATIONAL MARKETS We deliver the facts you make the decisions

Science Impact Enhancing the Use of USGS Science

TERMS OF REFERENCE FOR CONSULTANTS

Committee on Development and Intellectual Property (CDIP)

QLectives: evolving software to support quality

COUNTRY: Questionnaire. Contact person: Name: Position: Address:

Conducting National Biodiversity Assessments

and its repercussions on the scientific community Berlin, September 2013 Koenraad Debackere, KU Leuven

SFINNO Database of Finnish Innovations & Impact Assessment Via Innovations of the National STI Agency

International comparison of education systems: a European model? Paris, November 2008

Section 1: The Nature of Science

FACULTY SENATE ACTION TRANSMITTAL FORM TO THE CHANCELLOR

Social Innovation and new pathways to social changefirst insights from the global mapping

RECOMMENDATIONS. COMMISSION RECOMMENDATION (EU) 2018/790 of 25 April 2018 on access to and preservation of scientific information

E-Training on GDP Rebasing

Armenian Experience on Agricultural Census

RACE TO THE TOP: Integrating Foresight, Evaluation, and Survey Methods

Belgian Position Paper

Original: English Introduction to all things related to coastal fisheries and aquaculture data

Why we need a Network of Usage Data Providers - OpenAIRE Impact Metrics Results

Linking Science to Technology - Using Bibliographic References in Patents to Build Linkage Schemes

Research Content, Workflows and Beyond. Lim Kok Keng

The University of the Future - as Education for Sustainable Development Hub

A Bibliometric Analysis of Australia s International Research Collaboration in Science and Technology: Analytical Methods and Initial Findings

U252 - Environmental Law Monday and Wednesday 11:00 a.m. -12:20 p.m. in SSPA 1165

Prepared by the Working Group on the Use of Nuclear Power Sources in Outer Space

Commission on science and Technology for Development. Ninth Session Geneva, May2006

Studying the Role of Public Research Organisations

esss Berlin, 8 13 September 2013 Monday, 9 October 2013

STI OUTLOOK 2002 COUNTRY RESPONSE TO POLICY QUESTIONNAIRE CZECH REPUBLIC. 1. General framework and trends in science, technology and industry policy

OpenUP. IRCDL 2018 Udine, Gennaio

Elaboration of Smart City Indicators System as a Guideline and a Common Framework for Implementation of Smart City Initiatives in the APEC Region

Title: The Impact of Altmetrics on library for biomedical research

EUROPEAN MANUFACTURING SURVEY EMS

UNIVERSIDAD CARLOS III DE MADRID

Innovation-Based Economic Development Strategy for Holyoke and the Pioneer Valley

Recommendations of the Microgravity Review Panel

Measuring Romania s Creative Economy

SciELO SA: Past, Present and Future (September 2018)

THE TOP 100 CITIES PRIMED FOR SMART CITY INNOVATION

Violent Intent Modeling System

INTERNATIONAL TELECOMMUNICATION UNION

Review of the University vision, ambition and strategy January 2016 Sir David Bell KCB, Vice-Chancellor

Planning for the 2010 Population and Housing Census in Thailand

Experiences from the Social Sciences - possible links to Health Data?

Route Planning & Cable Route Surveys

De staat van de sociale wetenschap en hoe die te meten. Paul Wouters and Thed van Leeuwen 27 September, 2012

THE SUBJECT COMPOSITION OF THE WORLD'S SCIENTIFIC JOURNALS

2020 Census: A Look at How Data Provides Answers to Questions and How States Work With the Census Bureau to Ensure Quality Data

INTERNATIONAL ATOMIC ENERGY AGENCY 58TH GENERAL CONFERENCE (22 26 September 2014)

RESEARCH PROGRAMME MANAGER for the AntiMicrobial Resistance Benchmark

Determination of the Marginal Exploitation Lifespan of Transformers Based on a Comprehensive Diagnostic Survey

Economic and Social Council

Competent Persons and Experts in Russia. Stephen Henley and Grigoriy Malukhin CRIRSCO November 2011

E-URAL. European Union and RussiA Link for S&T cooperation in the area of the environment

Challenges of Library and Information Science Journals: editor's opinions A survey approach

Practical measures to encourage the diffusion of green technologies: Licensing Fast tracking of green patents The GreenXchange Platform

International Transaction Journal of Engineering, Management, & Applied Sciences & Technologies

Chemicals Risk Management and Critical Raw Materials

Project Administration Instructions

Information Communication Technology

OECD Science, Technology and Industry Outlook 2008: Highlights

II. BULGARIAN E-READINESS ASSESSMENT MODEL AND METHODOLOGY FOR QUANTITATIVE ASSESSMENT

Characterizing Research and Technology Organizations (RTOs) using bibliometrics - The case of the Netherlands

CRC Association Conference

FINLAND. The use of different types of policy instruments; and/or Attention or support given to particular S&T policy areas.

Revisiting the USPTO Concordance Between the U.S. Patent Classification and the Standard Industrial Classification Systems

ILNAS-EN 14136: /2004

Transmission Availability Data System Phase II Final Report

Transcription:

Research assessment and evaluation in Russian fundamental science Denis Kosyakov and Andrey Guskov State Public Scientifiс Technological Library of the Siberian Branch of the Russian Academy of Sciences, Novosibirsk CRIS2018 session 5: POLICY AND ASSESSMENT Umeå, Sweden - June 14

National research assessment exercises

National research assessment Country Program or Authority Year(s) UK Research Assessment Exercise (RAE) 1989, 1992, 1996, 2001, 2008 UK Research Excellence Framework (REF) 2014 Netherlands Standard Evaluation Protocol (SEP) 1993 France Comité National d Évaluation, CNE 1989 2006 France Agency for the Evaluation of Research and Higher Education, AERES 2007 - Italy Triennial Evaluation Exercise (VTR) 2006-2009 Italy Quality of Research Assessment (VQR) 2011-2014 And many others

Methods classification

Research assessment in Russia hard path milestones

Start 2009 Legislative framework 2010-2011 First round / no consequencies

Second round 2013 Legislative framework was revised 2014 Start of paperwork and first elements of infrastructure

Second round 2014 Quantitative assessment methodology 2016 Federal monitoring system

Second round 2016-2017 Quantitative evaluation 2017 Expert assessment

Federal monitoring system

Federal system for monitoring of the effectiveness of scientific organizations performing scientific research, experimental design and technological work Federal monitoring system (sciencemon.ru) 1781 organizations 78 indicators

Indicators 4 MAIN GROUPS the effectiveness and relevance of scientific research; development of human resources; integration into the world scientific space, dissemination of scientific knowledge and increasing the prestige of science; resource support for the research activities PROBLEMS AND ISSUES aggregated indicators: misunderstandings and interpretations hard to verify (and prove) possibility of manipulation

Quantitative assessment methodology

Basics Indicators: 19 / 78 3 main determinants and 2 supplemental 3 main activity profiles + 1 extra profile 39 fields of science 117 reference groups 3 performance categories

Stages Calculation of determinants Determination of reference group Calculation of benchmarks Calculation of performance category

Main determinants A B C The number of books or scientific periodicals and papers indexed in Web of Science, Scopus or other specialized database (maximum value from one of the databases representative for an organization) per 100 researchers (FTE). Number of IPRs registered in the Russian Federation or abroad as well as the number of issued design documentation per 100 researchers (FTE). Income from all types of R&D, S&T services provided per total R&D personnel of an organization.

Activity profiles Determinant A Knowledge generation Determinant B Technology development Determinant C Provision of S & T services

Category assignment Thresholds med = median of main determinants in the reference group min threshold = 0.75 * med max threshold = 1.25 * med Category 1. Leading institutions 2. Sustained institutions 3. Degraded institutions (loosing research functions)

Example

FASO of Russia departmental assessment exercise

Quantitative assessment was based on the Federal monitoring system data for 2013-2015 trial calculation by the general methodology showed unsatisfactory results adjustments were proposed final results were calculated

Main proposed changes The distribution of organizations by reference groups was made only by the main field of research. The main determinant C was calculated as a ratio of income from all types of R&D, S&T services to basic funding amount. The thresholds for activity profile Knowledge generation were calculated within reference groups, and for activity profiles Technology development, Provision of S & T services within all institutions, because the main determinants for these profiles are less biased by the main field of research. Categories were calculated for all three activity profiles simultaneously. The resulting profile and category of organization was chosen by selecting the best assigned category in all profiles.

Determinant A mean, median and dispersion in reference groups

Distribution of determinant A in the reference group Geology, geochemistry, mineralogy

number of publications (WoS) number of researchers All institutions, 2014.

Reference group Psychology and Pedagogical Sciences

Reference group Philosophy

Results of quantitative assessment 2013 2014 2015

Expert assessment Results of quantitative assessment Final evaluation Selfevaluation survey, experts

Comparison of quantitative stage results and final assessment assigned categories distributions Science field 2013-2015 average distribution by category Final distribution by category Cat 1 Cat 2 Cat 3 Cat 1 Cat 2 Cat 3 Natural Sciences Technical Sciences 29% 49% 22% 45% 45% 11% 36% 46% 18% 28% 48% 23% Social Studies 14% 44% 42% 20% 49% 31% Humanities 11% 49% 40% 56% 41% 4% Health Sciences Agricultural Sciences 30% 49% 20% 21% 59% 21% 18% 55% 27% 20% 43% 37%

Lessons and conclusions Pure theoretical approach to the evaluation methodology design is unlikely to lead to good results. Uneven distribution of organizations by reference groups makes assessment difficult and unreliable for small ones. It s hard to achieve first category for diversified institutions with strong technology component in quantitative evaluation. Usage of indicators relative to number of researchers in organization can be significantly biased by incorrect count of researchers. Questionable character of determining of total number of publications as maximum from several sources.

Recent novelties in FASO research assessment and reporting

Evaluation of publications quality Fractional authorship Journal quality : Quartile Score (SCI), indexing in Scopus, listing in Russian High Attestation Commission special register The new standard of quality DOI!!! It turned out that 4704 research projects (45% of all) do not have any articles with assigned DOI (Digital Object Identifier). This means that the information network does not contain data on these publications. I emphasize that there is no data not only in the Web of Science database, but in general nowhere, as DOI today must have all the more or less significant publications! It turns out that the performers of the good half have nothing to report Alexey Khokhlov, Vice President of RAS

Methodology of calculation of quality of publication activity of research organization M total number of publications Q Quartile of the journal (SCI) N number of article authors a number of authors affiliations T num i - number of authors affiliations with FASO organizations

Compliance with the statements of the Leiden Manifesto for Research Metrics

The Leiden Manifesto for research metrics Quantitative evaluation should support qualitative, expert assessment + Measure performance against the research missions of the institution + Protect excellence in locally relevant research - Keep data collection and analytical processes open, transparent and simple - Allow those evaluated to verify data and analysis - Account for variation by field in publication and citation practices + Avoid misplaced concreteness and false precision - Recognize the systemic effects of assessment and indicators - Scrutinize indicators regularly and update them -

Main source of spotted problems

Prerequisites Implementation of a project approach to planning and managing basic scientific research in accordance with the Programs of fundamental scientific research of the State academies of sciences for 2008-2012 and 2013-2020. Transition to the service-oriented model of the public administration, which was expressed in the introduction of the concept of the state assignment in 2007, which led to a change in the funding model of scientific organizations from 2012.

Design and implementation sequence The main cornerstones of the methodology were laid at the highest level of authority at the first stage without proper study and preparations. Independence of separate teams responsible for the main parts indicators and monitoring system, methodology (calculations and procedures) of assessment Lack of public discussions and expert community involvement.

Ministry of Finance is at the core of developed approaches: Single year planning Elementary indicators Straightforward way to the calculation of necessary funding (number of FTE * standard costs)

GOODHART S LAW Charles Albert Eric Goodhart «As soon as the government attempts to regulate any particular set of financial assets, these become unreliable as indicators of economic trends» In scientometrics: «All metrics of scientific evaluation are bound to be abused» Mario Biagioli Biagioli, Mario (12 July 2016). «Watch out for cheats in citation game». Nature. 535 (7611): 201. doi:10.1038/535201a

Denis Kosyakov kosyakov@spsl.nsc.ru State Public Scientific Technological Library of the Siberian Branch of the Russian Academy of Sciences