In the peer
review system, science is represented by two separate yet equally important
groups: the reviewers, who examine manuscripts, and the associate editors, who
publish the articles. These are their stories [1].
I was asked to review a manuscript recently, and as
I was reading it, I started to think that the authors had HARKed [2] about half
their hypotheses. Although I’ve reviewed manuscripts before, this was the first
time I caught the odor of questionable research practices (QRPs). In the
interest of preserving the confidentiality of the review process, I’m not going
to provide any details about the manuscript and will instead focus on my own
experience, which was confusing, somewhat distressing, and led me to consider
what reviewers should do when they suspect researchers didn’t do things quite
right.
When I became suspicious, the first thing I did was
search for guidelines on what to do when this happens. I did a few Google
searches for reviewer guidelines, and I checked the journal’s website for
policies. I didn’t find anything relevant. Admittedly, I probably could have
tried harder. You could justifiably take me to task for not exhaustively
looking for guidelines for how to handle this – but should it be so difficult
to find guidance for this [3]? Researchers review papers all the time. This can’t
be the first time someone has been suspicious about a manuscript.
Uncertain of what to do exactly, I reverted to my
most basic impulses: I kept notes on everything. I documented, as best I could,
the basis for my suspicions. The paper in which Norbert Kerr coined the term
HARKing was particularly useful, especially the section called in which he
describes some probable symptoms of HARKing. Indeed, the hypotheses in the
manuscript seemed to have all the symptoms: theoretical reasoning that could
have easily led to the opposite predictions, counter-intuitive predictions that
had little basis in past literature but were ostensibly supported, and poor fit
between the hypotheses and methods. But it was actually the title of the
section in Kerr’s paper that had the largest impact on my thinking: “Circumstantial
evidence” is what he called those symptoms.
My official job title is Senior Lecturer of Legal Psychology. I think about evidence
a lot. So the evidentiary term got me thinking. What kind of investigation was
I conducting? How far should I take this? What was my role in this process?
To the last question – I had no idea, really, what
my job was. I knew I had responsibilities, but I wasn’t entirely sure what they
were [4]. I fired off an email to the associate editor handling the manuscript and
asked about relevant policies. While I was waiting for a reply, two
contingencies occurred to me: (1) if the journal has a policy concerning QRPs
in general or HARKing in particular, my job is similar to that of an American
police investigator and the editor is like a prosecutor, and (2) if the journal
has no specific policy on QRPs, my job is similar to an inquisitorial
prosecutor and the editor is like a judge. The analogy is far from perfect, but
I think these models are potentially useful ways of thinking about the two
situations. I whimsically call Model (1) the Investigative Model and Model (2)
the Prosecutorial Model. Let me elaborate…
In the Investigative Model, there are established
editorial policies on QRPs. For example, perhaps a journal might have a policy
to reject manuscripts that report questionable research [5]. This constrains
the roles of both the reviewers and the editor. As I see it, editorial policy
overrides the discretionary power a reviewer might otherwise have if no such
policy existed. If a reviewer suspects QRPs, his or her job would be to
assemble evidence, circumstantial though it may be, concerning the occurrence of the suspected QRPs. That
is, it’s the reviewer’s principal job, as it pertains to this issue, to raise
the question of whether or not the QRPs occurred and provide evidence. In the
same way that American police cannot directly influence sentencing of convicted
defendants, the reviewer’s job isn’t necessarily to offer suggestions or
recommendations for how to deal with the QRPs, since the journal’s policies
should describe what is supposed to happen. It is then the editor’s job to
determine whether there’s a case for the occurrence of QRPs and, if so, follow
the journal’s policy (or possibly exercise some good old fashioned
prosecutorial discretion). Regardless of the editor’s decision, the reviewer’s
role is essentially bound to identifying QRPs, not adjudicating them [6].
In the Prosecutorial Model, there are no editorial
policies on QRPs. This is a much looser situation. Presumably, in this scenario,
QRPs would be considered under the kind of “methodological soundness” criterion
that most journals have – which is probably not very specific. Here, a
substantial amount of discretion is granted to both reviewers and to the editor,
but I think the reviewer’s responsibilities also expand here. If a reviewer
suspects QRPs in this model, his or her job is to assemble a case, present the
evidence, and offer a “sentencing recommendation” – that is, suggestions for
how to remedy the situation. A reviewer’s role is expanded here, in my opinion,
because it is not the editor’s job to specifically detect and address QRPs.
Since QRPs fall under the general umbrella of methodological soundness here, a
reviewer would have to “make a case” that the QRPs (1) might have happened and
(2) are methodological problems if they happened. The editor then acts like a
judge and adjudicates the case, deciding how to proceed with a substantial
amount of discretion. It’s also worth pointing out that reviewers have enormous
discretion here, too: If a reviewer doesn’t think HARKing is really a problem,
for example, even if he or she thinks the authors HARKed, the reviewer can use choose
not make a fuss about it. Like a prosecutor undercharging a cooperative
informant, a reviewer might, say, allow an especially “interesting” or “surprising”
set of results pass with lighter scrutiny. That could lead to trouble.
These models imply very different ways of writing a
review that raises the specter of QRPs. In the Investigative Model, I think it’s
outside the reviewer’s purview to get on a soapbox about what they think should
happen concerning the QRPs. In the Prosecutorial Model, it’s exactly the
reviewer’s job to get on that particular soapbox and offer an opinion. My
thinking about this is still inchoate, but my intuition is that the
Investigative Model is probably more effective in handling suspected QRPs in a
non-arbitrary, procedurally sound manner. However, there may be other
unintended negative consequences [7].
Returning to my story – when the editor got back to
me, she indicated that there was no editorial policy on HARKing. Thus, in the
framework I had spontaneously generated, I was working in a Prosecutorial
framework. I wasn’t exactly comfortable with the role, but I set about acting
as a responsible prosecutor: I noted my suspicions and their basis, strove not
to overstate the case, and offered recommendations for how the authors should
revise the manuscript to be more transparent if they did indeed HARK. I also
signed the review. There are several good reasons to sign reviews, and among
them is sense of accountability it can induce. That is, knowing I was putting
my name on it made me more careful about how I presented my thoughts – and that
seemed to me especially important if I was going to raise suspicions of QRPs.
Moreover, it simply seemed more civilized not to cast suspicions from the
shadow of anonymity.
Ultimately, I was comfortable with my review and the
recommendations therein, but I am less comfortable with the general ambiguity
of reviewers’ role as it pertains to QRPs. If this isn’t already a
conversation, it needs to be one. And if it is already a conversation, it needs
to be louder.
References
Kerr, N. L. (1998). HARKing:
Hypothesizing after the results are known. Personality and Social
Psychology Review, 2, 196-217.
1: Executive Producer: Dick Wolf
2: Hypothesized After the Results are Known, for the
uninitiated.
3: In retrospect, maybe I could have reached out to
open-science savvy researchers on social media for advice. That probably would
have been a good idea. I’m still getting used to public engagement. I suspect
someone has already written about this – and quite likely something better
thought-out than this. If you know of anything relevant, please send it to me
or leave it in the comments.
4: Beyond this particular problem, the role of a
reviewer is often fraught with ambiguities, largely because, I think, editorial
standards vary so widely.
5: Off the top of my head, I don’t know if any
journals actually have policies remotely similar to this. My intuition is that
such hardline policies would be extremely difficult to enforce.
6: Never mind for now how actually difficult it is to
detect QRPs. By definition, many of them are invisible. How do you detect
unreported variables? Maybe sometimes there are traces, but I’d guess that most
of the time, there’s nothing to find.
7: One possibility is that if some journals adopted
Investigative Model policies, researchers who engage in QRPs might be driven to
roll the dice with journals that use the Prosecutorial Model and hope reviewers
don’t catch their shenanigans.
No comments:
Post a Comment