Transparent Replications

by Clearer Thinking

Rapid replications for reliable research

Why We Introduced the Clarity Criterion for the Transparent Replications Project


Updated: May 7, 2026

At Transparent Replications we have introduced a study rating criterion that, to our knowledge, has not been used before. We call it our “clarity” rating, and it is an assessment of how unlikely it would be for a reader to misinterpret the results of the study based on the discussion in the paper. 

When a study replicates, it is natural to assume that the claims the paper makes based on that study are likely to be true. Unfortunately, this is not necessarily the case, as there may be a substantial gap between what a study actually demonstrated and what the paper claims was demonstrated. All a replication shows is that with new data you can get the same statistical result; it doesn’t show that the claims made based on the statistical result are correct. The clarity rating helps address this, by evaluating the size of the gap between what was shown and what was claimed.

It’s important that papers have a high level of “clarity.” When they don’t, readers may conclude that studies support conclusions that aren’t actually demonstrated by the tests that were conducted. This causes unproven claims to make their way into future research agendas, policymaking, and individual decision-making.

We acknowledge that making a paper clear is a difficult task, and we ourselves often have room for being clearer in our explanations of results. We also acknowledge that most authors strive to make their papers as clear and accurate as possible. A perfect “clarity” rating is very difficult to obtain, and when a paper loses points on this criterion, we are in no way assuming that the authors have intentionally introduced opportunities for misinterpretation – on the contrary, it seems to us that most potential misinterpretations are easier for an outside research team to see, and most original authorship teams would prefer to avoid misinterpretations of their results.  

We hope that evaluating clarity will also serve to detect and disincentivize Importance Hacking. Importance Hacking is a new term we’ve introduced to refer to when real, replicable findings are made to appear more valuable or worthy of publication than they really are. This can (and probably often does) happen unintentionally. In some cases, Importance Hacking can convince journal editors and reviewers that a paper merits publication when, in fact, it doesn’t. We refer to such papers as “Importance-Hacked acceptances.” Had no Importance Hacking occurred, the paper would not have been accepted by the journal, even though the finding is real and replicable.

A variety of types of Importance Hacking exist, and they can co-occur. They attempt to inflate the value of the paper in at least one of the following ways:

  • Hacking Rigor: making a finding, study design, or methodological approach appear more valid, robust, conclusive, unbiased, or diagnostic of a stated claim than it really is.
  • Hacking Novelty: making a finding or methodological approach appear more original, surprising, or contrary to current consensus than it really is. 
  • Hacking Relevance: making a topic or finding appear to have a broader scope of contribution than it really does, often by overstating its practical, translational, or theoretical value.
  • Hacking Beauty: making a finding or study design appear more elegant, attractive, clean, simple, or internally consistent than it really is.
  • Hacking Effort: making a study design appear more arduous, resource-intensive, or technically challenging than it really is, thereby inflating the perceived rarity or impressiveness of the resulting data

Given that there is limited attention, money, and time for scientific research, these Importance Hacked studies use up limited space, attention, and resources that could be directed to more valuable work.

While we believe that evaluating the clarity of papers is important, it does have the drawback that it is more subjective to evaluate than other criteria, such as replicability. We try to be as objective as possible by focusing first on whether a study’s results directly support the claims made in the paper about the meaning of those results. This approach brings into focus any gap between the results and the authors’ conclusions. We also consider the completeness of the discussion – if there were study results that would have meaningfully changed the interpretation of the findings, but that were left out of the paper’s discussion, that would lower the clarity rating.

When replicating studies, we aim to pre-register not only the original analyses, but also the simplest valid statistical test(s) that could address a paper’s research questions, even if these were not reported in the original paper. Sometimes more complex analyses obscure important information. In such cases, it is useful to report on the simple analyses so that we (and, importantly, our readers) can see the answer to the original research questions in the most straightforward way possible. If our redone analysis using the simplest valid method is consistent with the original result, that lends it further support.

We would encourage other projects aiming to evaluate the quality of papers to use our clarity criterion as well. Transparency and replicability are necessary, but not sufficient, for quality research. Even if a study has been conducted transparently and is replicable, this does not necessarily imply that the results are best interpreted in exactly the way that the original authors interpreted them.

To understand our entire ratings system, read more about our transparency and replicability ratings.

Note: This article was updated on May 7, 2026 to reflect a more refined definition of Importance Hacking.