Systematic Differences in Impact across Publication Tracks at PNAS

DSpace/Manakin Repository

Systematic Differences in Impact across Publication Tracks at PNAS

Citable link to this page


Title: Systematic Differences in Impact across Publication Tracks at PNAS
Author: Rand, David Gertler; Pfeiffer, Thomas

Note: Order does not necessarily reflect citation order of authors.

Citation: Rand, David G. and Thomas Pfeiffer. 2009. Systematic differences in impact across publication tracks at PNAS. PLoS ONE 4:e8092.
Full Text & Related Files:
Abstract: Background: Citation data can be used to evaluate the editorial policies and procedures of scientific journals. Here we investigate citation counts for the three different publication tracks of the Proceedings of the National Academy of Sciences of the United States of America (PNAS). This analysis explores the consequences of differences in editor and referee selection, while controlling for the prestige of the journal in which the papers appear. Methodology/Principal Findings: We find that papers authored and “Contributed” by NAS members (Track III) are on average cited less often than papers that are “Communicated” for others by NAS members (Track I) or submitted directly via the standard peer review process (Track II). However, we also find that the variance in the citation count of Contributed papers, and to a lesser extent Communicated papers, is larger than for direct submissions. Therefore when examining the 10% most-cited papers from each track, Contributed papers receive the most citations, followed by Communicated papers, while Direct submissions receive the least citations. Conclusion/Significance: Our findings suggest that PNAS “Contributed” papers, in which NAS–member authors select their own reviewers, balance an overall lower impact with an increased probability of publishing exceptional papers. This analysis demonstrates that different editorial procedures are associated with different levels of impact, even within the same prominent journal, and raises interesting questions about the most appropriate metrics for judging an editorial policy's success.
Published Version: doi:10.1371/journal.pone.0008092
Other Sources:
Terms of Use: This article is made available under the terms and conditions applicable to Other Posted Material, as set forth at
Citable link to this page:
Downloads of this work:

Show full Dublin Core record

This item appears in the following Collection(s)


Search DASH

Advanced Search