For Peer Review Week 2016, we set a simple competition for you all, to publicly peer review one of 25 million research articles on our platform. This fitted perfectly with the theme this year of ‘Recognising Review’, as every single peer review conducted with us is published openly and creditable through the application of a CC BY license, which enables the unrestricted sharing and re-use of the reviews providing that attribution is given.
We’re happy to announce that Lauren Collister was the winner this year, and a t-shirt is on your way!
To honor and celebrate peer review, a group of organizations is working collaboratively to plan a week of activities and events. The group is delighted to announce that the second annual Peer Review Week will run from September 19- 25, 2016.
This year’s theme is Recognition for Review, exploring all aspects of how those participating in review activity – in publishing, grant review, conference submissions, promotion and tenure, and more – should be recognized for their contribution.
NOTE: OpenAIRE would like to know what you think about open peer review! Have your say here until 7th October!
Tl;dr – “Post-publication peer review” (PPPR) has gained a lot of traction in recent years. As with much of peer review’s confusing lexicon, however, this term is ambiguous. This ambiguity stems from confusion over what constitutes “publication” in the digital age. PPPR conflates two distinct phenomena, which we would do better to treat separately, namely “open pre-review manuscripts” and “open final-version commenting”.
What is “post-publication peer review”?
Peer review can have two senses, one specific and the other more general. “Peer Review” (henceforth PR) is a well-defined publishing practice for the quality assurance of research articles and other academic outputs. It is intimately tied to the publication process. It traditionally begins when an editor sends a manuscript to reviewers and ends when the editor accepts a manuscript for publication. But “peer review” (lower-case, henceforth “pr”) is just the critique and appraisal of ideas, theories, and findings by those with particular insight into a topic. Such feedback happens all the time. It happens before manuscripts are submitted: in colleagues’ initial reactions (positive or negative) to a new idea, feedback gained from conferences, lectures, seminars and late-night bull sessions, or private comments on late-stage first-draft manuscripts from trusted peers. And it continues after the article’s appearance in a journal, via a multitude of channels through which readers can give feedback, including comment sections on journal websites, dedicated channels for post-publication commentary, blogs and social media, and of course in future research that cites and comments back on the findings.
It’s the most wondrous time of the year! Peer Review Week is the time when the scholarly communications community comes together to recognise the importance and value of peer review and peer reviewers.
This year, the theme is all about Recognising Review, and the valiant efforts of the research community in performing and managing peer review.
To celebrate this, and discuss the topic further, we will be holding a webinar on September 19th at 11AM GMT all about new and future approaches for acknowledging the peer review process.
Peer review week is almost upon us, yippee! Taking place from the 19-26th of September, this is a global event for celebrating the critical role that peer review plays in the scholarly communication process.
The theme this week is recognising peer review, something that at ScienceOpen we’ve been passionate about since our inception. Other science startups like Publons have already launched amazing initiatives like their Sentinels of Science campaign, well worth checking out. For updates, follow the #RecognizeReview hashtag on Twitter!
This theme couldn’t be better timed. At the moment in scholarly publishing and communications, there’s something of a renaissance happening where we’re exploring aspects of recognising all the different types of review activity, including in publishing, research grants, conference submissions, and promotion and tenure.
There are many many amazing blogs and bloggers out there that provide critical comments, context, and feedback on the ‘formally published’ research literature. One problem with these though is that they are often divorced from the papers themselves, perhaps lost on obscure websites, or not hitting the right target audience. This seems like an awful waste, don’t you think?
While some great initiatives such as The Winnower will now publish blog posts openly, these still are not connected to the papers that they are based on, if they are indeed written about particular papers. But what do researchers think about blogging as a form of scholarly communication in the form of post-publication peer review?
So as with most of my ponderings, I took to Twitter to get some feedback with a little poll. I actually framed the question a little ambiguously, but this shouldn’t sufficiently skew the data in any direction (I hope).
Do you consider blogging to be a form of post-publication peer review?
What is interesting to me is that 41% of people who answered, who undoubtedly did not constitute just a researcher sample, do not consider blogging to ‘count’ as peer review. I would really love to know why this is the case for some people. Perhaps they haven’t seen good examples, or perhaps just because it’s not formalised in any way, and quite disassociated from the research literature.
But did you know that anyone can review any article they want on ScienceOpen, and not just those from ScienceOpen Research? And perhaps more importantly, anyone can invite anyone else to review any article? That sounds an awful lot like the daytime job for Editors at traditional journals.. But with the power firmly in the hand of researchers and their communities. How cool is that?
It’s super easy to implement too. All you have to do is go to an article of choice, click the ‘Reviews’ button (Step 1), and then select the ‘Invite to Review’ button (Step 2). If you were feeling inclined, you could review the paper yourself too!
You can then simply select their ScienceOpen username (what, you don’t have one yet?!), or invite them by email (Step 3).
I remember my first peer review. An Editor for a well-respected Elsevier journal in Earth Sciences emailed me during the second year of my PhD, asking me to peer review a paper for them. I hadn’t published anything by this point of my PhD, and had received no formal training in how to peer review papers. I initially declined, but was pretty much coerced into doing it, despite my resignations. “It’ll be great training and experience”, I was told. Go on. Go on go on go on go on go on. In the end, I did the review, but got my supervisor to check it over to make sure I was fair, thorough, and constructive. I remember him saying “This is surprisingly good!”, and thinking ‘Thanks..’. But his response was more because it was my first peer review, without any training in how to do it, rather than anything to do with my ability as a scientist. And rightly so – why should I have been expected to do a good job of peer review at such an early stage in my career, and with no formal training?
I wonder then how many other PhD students are told the same, and thrown into the deep end. ‘Peer review for this journal and receive fame and glory. It doesn’t matter how well you do it, as long as you do it.’
I get the feeling that some researchers regard public, post-publication peer review as a non-rigorous, non-structured and poor alternative to traditional peer review. Much of this might be down to the view that there are no standards, and no control in a world of ‘open’.
This couldn’t be further from the truth.
At venues like ScienceOpen and F1000 Research, there is full Editorial control over peer review. The only difference is that there is an additional safe guard against fraud and abuse. In public peer reviews, the quality (and quantity) of the process is made explicit. Both the report and the identity of the reporter are made open. This type of system invites civility and community engagement, and lays the foundation for crediting referees. It also highlights an under-appreciated, overlooked, aspect of the work that scientists do to advance knowledge in the real world.
ScienceOpen Editor Dan Cook said “Personally, I think the public needs to know how hard scientists work to advance our understanding of the world. “
At ScienceOpen, the Editorial office plays two roles. First, the Editorial team for ScienceOpen Research performs all the basic standards checks to make sure that research published is at an appropriate scientific standard. They attempt to protect against pseudoscience, and ensure that the manuscript is prepared to undergo public scrutiny. Second, there are Collection Editors, who manage peer review, curation, and discussion about their own Collections.
Why is Editorial control so important?
For starters, without an Editor, peer review will never get done. Researchers are busy, easily distracted, and working on 1000 other things at once. Opting to go out into the world and randomly distribute your knowledge through peer review, while selfless, is actually quite a rare phenomenon.
Peer review needs structure, coordination, and control. In the same way as traditional peer review, this can be facilitated by an Editor.
But why should this imply a closed system? In a closed system, who is peer reviewing the Editors? What are editorial decisions based on? Why and who are Editors selecting as reviewers?
These are all questions that are obscured by traditional peer review, and traits of a closed, secretive, and subjective system – not the rigorous, objective, gold standard that we hold peer review to be.
At ScienceOpen, we recognise this dual need for Editorial standards combined with transparency. Transparency leads to accountability, which in turn lends itself to a less biased, more rigorous and civil process of peer review.
How does Editorial coordination work with Collections?
Collections are the perfect place to demonstrate and exercise editorial management. Collection Editors, of which there can be up to five per Collection, have the authority to manage the process of peer review, but out in the open.
They can do this by either externally inviting colleagues to review papers within the system, or if they already have a profile with us, then they can simply invite them to review specific papers, and referees will receive an invitation to peer review.
Quality control is facilitated through ORCID, as referees must have 5 items associated with their account in order to formally peer review. And to comment, all you need is an ORCID account, simples!
The major difference between a traditional Editor and a Collection Editor is selection. As a traditional Editor, you wield supreme power over what ultimately becomes published in the journal by deciding what gets rejected and what gets sent out to peer review. As a Collection Editor, you don’t reject anything – you filter from pre-existing content depending on your scope.
The arXiv is a server that hosts ‘eprints’ or ‘preprints’ of research papers, and is a key publishing platform for many fields, particularly physics and mathematics. Founded back in 1991 by Paul Ginsparg, it currently hosts over 1 million research articles, with more than 8000 submissions per month!
Despite now being in the running for 25 years, the arXiv still represents one of the greatest technological innovations to utilise the Web for scholarly communication.
While the majority of the content submitted to the arXiv is subsequently also submitted to traditional journals for publication, there is still content which never goes beyond its confines. Irrespective of this, communities engaged with the arXiv still cite articles published there, whether or not they have been formally published in a journal elsewhere.
This is the whole purpose of the arXiv: to facilitate rapid peer-to-peer communication so that science accelerates faster. The fact that all articles are publicly available is incidental, and just happens to be a topic of major interest with the growing open access movement.
However, the arXiv is not peer reviewed in the formal sense. It is moderated, so that junk submissions can be removed, or manuscripts recategorised, but it lacks the additional layer of quality control of traditional peer review.
So while some might think this poses a risk, ask yourself this question: do you re-use articles critical to your research without making sure that you have checked and understand the research to a sufficient degree that you can appropriately cite it? Because that’s peer review, that is, and it applies irrespective of whether an article has already been peer reviewed or not.