The Influence of Peer Reviewer Expertise on the Evaluation of Research Funding Applications.

Although the scientific peer review process is crucial to distributing research investments, little has been reported about the decision-making processes used by reviewers. One key attribute likely to be important for decision-making is reviewer expertise. Recent data from an experimental blinded re...

Full description

Bibliographic Details
Main Authors: Stephen A Gallo, Joanne H Sullivan, Scott R Glisson
Format: Article
Language:English
Published: Public Library of Science (PLoS) 2016-01-01
Series:PLoS ONE
Online Access:https://doi.org/10.1371/journal.pone.0165147
id doaj-aa8d1a473bbc4059918fec99d569c07d
record_format Article
spelling doaj-aa8d1a473bbc4059918fec99d569c07d2021-03-03T20:32:55ZengPublic Library of Science (PLoS)PLoS ONE1932-62032016-01-011110e016514710.1371/journal.pone.0165147The Influence of Peer Reviewer Expertise on the Evaluation of Research Funding Applications.Stephen A GalloJoanne H SullivanScott R GlissonAlthough the scientific peer review process is crucial to distributing research investments, little has been reported about the decision-making processes used by reviewers. One key attribute likely to be important for decision-making is reviewer expertise. Recent data from an experimental blinded review utilizing a direct measure of expertise has found that closer intellectual distances between applicant and reviewer lead to harsher evaluations, possibly suggesting that information is differentially sampled across subject-matter expertise levels and across information type (e.g. strengths or weaknesses). However, social and professional networks have been suggested to play a role in reviewer scoring. In an effort to test whether this result can be replicated in a real-world unblinded study utilizing self-assessed reviewer expertise, we conducted a retrospective multi-level regression analysis of 1,450 individual unblinded evaluations of 725 biomedical research funding applications by 1,044 reviewers. Despite the large variability in the scoring data, the results are largely confirmatory of work from blinded reviews, by which a linear relationship between reviewer expertise and their evaluations was observed-reviewers with higher levels of self-assessed expertise tended to be harsher in their evaluations. However, we also found that reviewer and applicant seniority could influence this relationship, suggesting social networks could have subtle influences on reviewer scoring. Overall, these results highlight the need to explore how reviewers utilize their expertise to gather and weight information from the application in making their evaluations.https://doi.org/10.1371/journal.pone.0165147
collection DOAJ
language English
format Article
sources DOAJ
author Stephen A Gallo
Joanne H Sullivan
Scott R Glisson
spellingShingle Stephen A Gallo
Joanne H Sullivan
Scott R Glisson
The Influence of Peer Reviewer Expertise on the Evaluation of Research Funding Applications.
PLoS ONE
author_facet Stephen A Gallo
Joanne H Sullivan
Scott R Glisson
author_sort Stephen A Gallo
title The Influence of Peer Reviewer Expertise on the Evaluation of Research Funding Applications.
title_short The Influence of Peer Reviewer Expertise on the Evaluation of Research Funding Applications.
title_full The Influence of Peer Reviewer Expertise on the Evaluation of Research Funding Applications.
title_fullStr The Influence of Peer Reviewer Expertise on the Evaluation of Research Funding Applications.
title_full_unstemmed The Influence of Peer Reviewer Expertise on the Evaluation of Research Funding Applications.
title_sort influence of peer reviewer expertise on the evaluation of research funding applications.
publisher Public Library of Science (PLoS)
series PLoS ONE
issn 1932-6203
publishDate 2016-01-01
description Although the scientific peer review process is crucial to distributing research investments, little has been reported about the decision-making processes used by reviewers. One key attribute likely to be important for decision-making is reviewer expertise. Recent data from an experimental blinded review utilizing a direct measure of expertise has found that closer intellectual distances between applicant and reviewer lead to harsher evaluations, possibly suggesting that information is differentially sampled across subject-matter expertise levels and across information type (e.g. strengths or weaknesses). However, social and professional networks have been suggested to play a role in reviewer scoring. In an effort to test whether this result can be replicated in a real-world unblinded study utilizing self-assessed reviewer expertise, we conducted a retrospective multi-level regression analysis of 1,450 individual unblinded evaluations of 725 biomedical research funding applications by 1,044 reviewers. Despite the large variability in the scoring data, the results are largely confirmatory of work from blinded reviews, by which a linear relationship between reviewer expertise and their evaluations was observed-reviewers with higher levels of self-assessed expertise tended to be harsher in their evaluations. However, we also found that reviewer and applicant seniority could influence this relationship, suggesting social networks could have subtle influences on reviewer scoring. Overall, these results highlight the need to explore how reviewers utilize their expertise to gather and weight information from the application in making their evaluations.
url https://doi.org/10.1371/journal.pone.0165147
work_keys_str_mv AT stephenagallo theinfluenceofpeerreviewerexpertiseontheevaluationofresearchfundingapplications
AT joannehsullivan theinfluenceofpeerreviewerexpertiseontheevaluationofresearchfundingapplications
AT scottrglisson theinfluenceofpeerreviewerexpertiseontheevaluationofresearchfundingapplications
AT stephenagallo influenceofpeerreviewerexpertiseontheevaluationofresearchfundingapplications
AT joannehsullivan influenceofpeerreviewerexpertiseontheevaluationofresearchfundingapplications
AT scottrglisson influenceofpeerreviewerexpertiseontheevaluationofresearchfundingapplications
_version_ 1714821917176758272