Professional Documents
Culture Documents
Timothy Hugh Barker1 Jennifer C. Stone1 Kim Sears2 Miloslav Klugar3 Catalin Tufanaru4 Jo Leonardi-Bee5
1AWnYQp/IlQrHD3i3D0OdRyi7TvSFl4Cf3VC1y0abggQZXdtwnfKZBYtws= on 11/04/2023
1 1
Edoardo Aromataris Zachary Munn
1
JBI, Faculty of Health and Medical Sciences, The University of Adelaide, Adelaide, SA, Australia, 2Queen’s Collaboration for Health Care Quality,
Queen’s University, Kingston, ON, Canada, 3Czech National Centre for Evidence-Based Healthcare and Knowledge Translation (Cochrane Czech
Republic, The Czech Republic (Middle European) Centre for Evidence-Based Healthcare: A JBI Centre of Excellence, Masaryk University GRADE
Centre), Faculty of Medicine, Institute of Biostatistics and Analyses, Masaryk University, Brno, Czech Republic, 4Centre for Health Informatics,
Australian Institute of Health Innovation, Macquarie University, Sydney, NSW, Australia, and 5The Nottingham Centre for Evidence Based
Healthcare: A JBI Centre of Excellence, School of Medicine, University of Nottingham, Nottingham, UK
ABSTRACT
JBI recently began the process of updating and revising its suite of critical appraisal tools to ensure that these tools
remain compatible with recent developments within risk of bias science. Following a rigorous development process
led by the JBI Effectiveness Methodology Group, this paper presents the revised critical appraisal tool for the
assessment of risk of bias for randomized controlled trials. This paper also presents practical guidance on how the
questions of this tool are to be interpreted and applied by systematic reviewers, while providing topical examples.
We also discuss the major changes made to this tool compared to the previous version and justification for why
these changes facilitate best-practice methodologies in this field.
Keywords: critical appraisal tool; methodological quality; methodology; randomized controlled trial; risk of bias
JBI Evid Synth 2023; 21(3):494–506.
which may be different within the same study de- How to use the revised tool
Downloaded from http://journals.lww.com/jbisrir by BhDMf5ePHKav1zEoum1tQfN4a+kJLhEZgbsIHo4XMi0hCywCX
Reviewers may face cases where there are fewer than recommend that studies be removed from a system-
Downloaded from http://journals.lww.com/jbisrir by BhDMf5ePHKav1zEoum1tQfN4a+kJLhEZgbsIHo4XMi0hCywCX
7 outcomes being appraised for a particular RCT, atic review following critical appraisal.
and there are more than 3 results being appraised per By removing studies, it presupposes that the pur-
outcome. The tool can be edited as required by the pose of a systematic review is to only permit the
1AWnYQp/IlQrHD3i3D0OdRyi7TvSFl4Cf3VC1y0abggQZXdtwnfKZBYtws= on 11/04/2023
review team to facilitate their use in these cases. synthesis of high-quality studies. While it may readily
For example, consider a hypothetical RCT that has promote alignment to the best available evidence, it
included 2 outcomes relevant to a question of a limits the full potential of the processes of evidence
systematic review team. These outcomes are mortal- synthesis to fully investigate eligible studies, their
ity and quality of life, both of which have been data, and provide a complete view of the evidence
measured at 2 time points within the study. When available to inform the review question.14,15 There are
using this tool, questions 1–6 and 13 are universal to several other approaches to incorporate the results
both outcomes, as they are addressed at the study of critical appraisal into the systematic review or
level and are only answered once. The reviewer meta-analysis. These approaches can include meta-
should then address question 7–9 twice, once for each regression, elicitation of expert opinion, using prior
outcome that is being appraised. Likewise, questions distributions, and quality-effect modeling.16 How-
10–12 should be addressed separately for both out- ever, these techniques demand appropriate statistical
comes but should also be assessed for each of the expertise and are beyond the scope of this paper.
results that has contributed data toward that out- Regardless of the approach ultimately chosen by the
come (eg, mortality at time point 1 and 2). In this reviewers, the results of the critical appraisal process
example, the reviewer would assess questions 10–12 should always be considered in the analysis and inter-
4 different times. It is also important to note that, as pretation of the findings of the synthesis.
with other critical appraisal tools,3,13 this tool should
also be applied in duplicate and independently during Overall assessment and presentation of results
the systematic review process. Reviewers should also Previous iterations of the JBI critical appraisal tool
be wary to only appraise outcomes that are relevant for RCTs intuitively supported reviewers assessing
to their systematic review question. If the only rele- the overall quality of a study using a checklist-based
vant outcome from this RCT for the systematic or scale-based tool structure (each item can be quan-
review question was mortality, then appraising the tified, which is enumerated to provide an overall
outcome quality of life would not be expected. quality score).8 The revised tool has been designed
to also facilitate judgments specific to the domains of
Interpretation of critical appraisal bias in which the questions belong
Some reviewers may take the approach of removing A reviewer may determine that for study 1, there
studies from progressing to data extraction or syn- was a low risk of bias for the domain “selection and
thesis in their review following the critical appraisal allocation,” as all questions received a response of
process. Removal of a study following critical ap- “yes.” However, for study 2, a reviewer may deter-
praisal may involve considering whether a certain mine a moderate risk of bias for the same domain, as
criterion had not been met (eg, randomization not the response to one of the questions was “no.” Im-
being demonstrated may warrant removal, assuming portantly, we provide no thresholds for grading of
the review was not also including other study de- bias severity (ie, low, moderate, high, critical, or other
signs with lesser internal validity due to not attempt- approaches) and leave this to the discretion of the
ing randomization). Another approach may include user and specific context in which they are working.
the review team weighting each question of the tool Considering the questions and assessments in this
(eg, randomization may be twice as important as regard, looking across all included studies (or a single
blinding of the outcome assessors); if a study fails study) can permit the reviewer to readily comment on
to meet a predetermined weight (decided by the how the risk of bias may impact on the certainty of
review team), then it may be removed. Other ap- their results at this domain level in the GRADE ap-
proaches may be to use simple cutoff scores (eg, if a proach. A judgment-based approach as described
study is scored with 10 “yes” responses, then it is here is one way for users to adopt the revised JBI
included) or to exclude studies that have been judged critical appraisal tool for RCTs; however, the tool is
as having a high risk of bias.8 However, we do not still compatible with either a checklist-based or
scale-based structure,8 and the decision of which ap- themselves. These known characteristics of the parti-
Downloaded from http://journals.lww.com/jbisrir by BhDMf5ePHKav1zEoum1tQfN4a+kJLhEZgbsIHo4XMi0hCywCX
proach to follow is left to the discretion of the cipants may distort the comparability of the groups
review team. (ie, does the intervention group contain more people
Current tools to appraise RCTs5 ask the reviewer over the age of 65 compared to the control?). A true
1AWnYQp/IlQrHD3i3D0OdRyi7TvSFl4Cf3VC1y0abggQZXdtwnfKZBYtws= on 11/04/2023
to establish an overall assessment of the risk of bias random assignment of participants to the groups
for each appraised study and for the overall body of means that a procedure is used that allocates the
evidence (ie, all appraised studies). The revised JBI participants to groups purely based on chance, not
critical appraisal tool for RCTs does not strictly influenced by any known characteristics of the parti-
prescribe to this, regardless of the approach fol- cipants. Reviewers should check the details about the
lowed. However, if reviewers opt to establish an randomization procedure used for allocation of the
overall assessment, then these assessments should participants to study groups. Was a true chance (ran-
not take into consideration the questions regarding dom) procedure used? For example, was a list of
statistical conclusion validity (questions 11–13) be- random numbers used? Was a computer-generated
cause risk of bias is only impacted by the internal list of random numbers used? Was a statistician,
validity construct.8 external to the research team, consulted for the
Irrespective of the approach taken, the results of randomization sequence generation? Additionally,
critical appraisal using the revised JBI critical apprai- reviewers should check that the authors are not stat-
sal tool for RCTs should be reported narratively in ing they have used random approaches when they
the review. This narrative summary should provide have instead used systematic approaches (such as
both an overall description of the methodological allocating by days of the week).
quality and risk of bias at the domain level of the
Question 2: Was allocation to groups concealed?
included studies. There should also be a statement
Category: Internal validity
regarding any important or interesting deviations
Domain: Bias related to selection and allocation
from the observed trends. This narrative summary
Appraisal: Study level
can be supported with the use of a table or graphic
If those allocating participants to the compared
that shows how each included study responded. We
groups are aware of which group is next in the
recommend presenting the results of critical appraisal
allocation process (ie, the treatment or control
for all questions via a table rather than summarizing
group), there is a risk that they may deliberately
with a score; see the example in Table 1. (Note that
and purposefully intervene in the allocation of pa-
this design is not prescriptive and only serves as an
tients. This may result in the preferential allocation
example).
of patients to the treatment group or to the control
group. This may directly distort the results of the
The revised JBI critical appraisal tool for study, as participants no longer have an equal and
randomized controlled trials random chance to belong to each group. Conceal-
The criteria and considerations that should be made ment of allocation refers to procedures that prevent
by reviewers when answering the questions in the those allocating patients from knowing before allo-
revised JBI critical appraisal tool for RCTs are cation which treatment or control is next in the
shown in Table 2. This tool is also available to allocation process. Reviewers should check the de-
download as Supplemental Digital Content 1 at tails about the procedure used for allocation con-
http://links.lww.com/SRX/A7. cealment. Was an appropriate allocation conceal-
ment procedure used? For example, was central
Question 1: Was true randomization used for assign-
randomization used? Were sequentially numbered,
ment of participants to treatment groups?
opaque, and sealed envelopes used? Were coded
Category: Internal validity
drug packs used?
Domain: Bias related to selection and allocation
Appraisal: Study level Question 3: Were treatment groups similar at the
If participants are not allocated to treatment and baseline?
control groups by random assignment, there is a risk Category: Internal validity
that this assignment to groups can be influenced by Domain: Bias related to selection and allocation
the known characteristics of the participants Appraisal: Study level
METHODOLOGY
© 2023 JBI. Unauthorized reproduction of this article is prohibited.
Table 1: Example presentation of results following critical appraisal using the revised JBI critical appraisal tool for randomized
controlled trials
QUESTION NO. 1 2 3 4 5 6 7 8 9 10 11 12 13
known characteristics of participants included in the Like question 4, those delivering the treatment who
compared groups constitutes a threat to internal va- are aware of participant allocation to either treat-
lidity. If differences in these characteristics do ment or control may treat participants differently
1AWnYQp/IlQrHD3i3D0OdRyi7TvSFl4Cf3VC1y0abggQZXdtwnfKZBYtws= on 11/04/2023
exist, then there is potential that the effect cannot compared to those who remain unaware of parti-
be attributed to the potential cause (the examined cipant allocation. There is a risk that any potential
intervention or treatment). This is because the effect change in behavior may influence the implementa-
may be explained by the differences between partici- tion of the compared treatments, and the results of
pant characteristics and not the intervention/treat- the study may be distorted. Blinding of those deli-
ment of interest. Reviewers should check the charac- vering treatment is used to minimize this risk.
teristics reported for participants. Are the participants When this level of blinding has been achieved,
from the compared groups similar with regard to the those delivering the treatment are not aware if they
characteristics that may explain the effect, even in the are treating the group receiving the treatment of
absence of the cause (eg, age, severity of the disease, interest or if they are treating any other group
stage of the disease, coexisting conditions)? Re- receiving the control intervention. Reviewers
viewers should check the proportion of participants should check the details reported in the article
with specific relevant characteristics in the compared about the blinding of those delivering treatment
groups. (Note: Do not only consider the P value for with regard to treatment assignment. Is there any
the statistical testing of the differences between
information in the article about those delivering
groups with regard to the baseline characteristics.)
the treatment? Were those delivering the treatment
Question 4: Were participants blind to treatment unaware of the assignments of participants to the
assignment? compared groups?
Category: Internal validity
Question 6: Were treatment groups treated identi-
Domain: Bias related to administration of interven-
cally other than the intervention of interest?
tion/exposure
Category: Internal validity
Appraisal: Study level
Domain: Bias related to administration of interven-
Participants who are aware of their allocation to
tion/exposure
either the treatment or the control may behave, re-
Appraisal: Study level
spond, or react differently to their assigned treatment
To attribute the effect to the cause (assuming there
(or control) compared with participants who remain
is no bias related to selection and allocation), there
unaware of their allocation. Blinding of participants
should be no difference between the groups in
is a technique used to minimize this risk. Blinding
terms of treatment or care received, other than
refers to procedures that prevent participants from
the treatment or intervention controlled by the
knowing which group they are allocated. If blinding
researchers. If there are other exposures or treat-
has been followed, participants are not aware if they
ments occurring at the same time as the cause (the
are in the group receiving the treatment of interest or
treatment or intervention of interest), then the ef-
if they are in another group receiving the control
fect can potentially be attributed to something
intervention. Reviewers should check the details re-
other than the examined cause (the investigated
ported in the article about the blinding of participants
treatment). This is because it is plausible that the
with regard to treatment assignment. Was an appro-
effect may be explained by other exposures or
priate blinding procedure used? For example, were
treatments that occurred at the same time as the
identical capsules or syringes used? Were identical
cause. Reviewers should check the reported expo-
devices used? Be aware of different terms used; blind-
sures or interventions received by the compared
ing is sometimes also called masking.
groups. Are there other exposures or treatments
Question 5: Were those delivering the treatment occurring at the same time as the cause? Is it
blind to treatment assignment? plausible that the effect may be explained by other
Category: Internal validity exposures or treatments occurring at the same time
Domain: Bias related to administration of interven- as the cause? Is it clear that there is no other
tion/exposure difference between the groups in terms of treatment
or care received, other than the treatment or inter- statistical relationship between the cause and the
Downloaded from http://journals.lww.com/jbisrir by BhDMf5ePHKav1zEoum1tQfN4a+kJLhEZgbsIHo4XMi0hCywCX
to know the number of participants and the propor- Were participants analyzed in the groups to which
Downloaded from http://journals.lww.com/jbisrir by BhDMf5ePHKav1zEoum1tQfN4a+kJLhEZgbsIHo4XMi0hCywCX
tions of participants with incomplete data; the rea- they were initially randomized, regardless of
sons for loss to follow-up are essential in the analysis whether they participated in those groups and
of risk of bias. regardless of whether they received the planned
1AWnYQp/IlQrHD3i3D0OdRyi7TvSFl4Cf3VC1y0abggQZXdtwnfKZBYtws= on 11/04/2023
Choice - comments/
Internal validity justification Yes No Unclear N/A
Table 2: (continued )
Downloaded from http://journals.lww.com/jbisrir by BhDMf5ePHKav1zEoum1tQfN4a+kJLhEZgbsIHo4XMi0hCywCX
Choice - comments/
Internal validity justification Yes No Unclear N/A
1AWnYQp/IlQrHD3i3D0OdRyi7TvSFl4Cf3VC1y0abggQZXdtwnfKZBYtws= on 11/04/2023
Choice - comments/
Statistical conclusion validity justification Yes No Unclear N/A
Table 2: (continued )
Downloaded from http://journals.lww.com/jbisrir by BhDMf5ePHKav1zEoum1tQfN4a+kJLhEZgbsIHo4XMi0hCywCX
Choice - comments/
Statistical conclusion validity justification Yes No Unclear N/A
1AWnYQp/IlQrHD3i3D0OdRyi7TvSFl4Cf3VC1y0abggQZXdtwnfKZBYtws= on 11/04/2023
Result 2 □ □ □ □
Result 3 □ □ □ □
Outcome 4 Yes No Unclear N/A
Result 1 □ □ □ □
Result 2 □ □ □ □
Result 3 □ □ □ □
Outcome 5 Yes No Unclear N/A
Result 1 □ □ □ □
Result 2 □ □ □ □
Result 3 □ □ □ □
Outcome 6 Yes No Unclear N/A
Result 1 □ □ □ □
Result 2 □ □ □ □
Result 3 □ □ □ □
Outcome 7 Yes No Unclear N/A
Result 1 □ □ □ □
Result 2 □ □ □ □
Result 3 □ □ □ □
12 Was appropriate statistical analysis used?
Outcome 1 Yes No Unclear N/A
Result 1 □ □ □ □
Result 2 □ □ □ □
Result 3 □ □ □ □
Outcome 2 Yes No Unclear N/A
Result 1 □ □ □ □
Result 2 □ □ □ □
Result 3 □ □ □ □
Outcome 3 Yes No Unclear N/A
Result 1 □ □ □ □
Result 2 □ □ □ □
Result 3 □ □ □ □
Outcome 4 Yes No Unclear N/A
Result 1 □ □ □ □
Result 2 □ □ □ □
Result 3 □ □ □ □
Outcome 5 Yes No Unclear N/A
Result 1 □ □ □ □
Result 2 □ □ □ □
Result 3 □ □ □ □
Outcome 6 Yes No Unclear N/A
Result 1 □ □ □ □
Result 2 □ □ □ □
Result 3 □ □ □ □
Table 2: (continued )
Downloaded from http://journals.lww.com/jbisrir by BhDMf5ePHKav1zEoum1tQfN4a+kJLhEZgbsIHo4XMi0hCywCX
Choice - comments/
Statistical conclusion validity justification Yes No Unclear N/A
1AWnYQp/IlQrHD3i3D0OdRyi7TvSFl4Cf3VC1y0abggQZXdtwnfKZBYtws= on 11/04/2023
symptoms). Crossover trials should ensure there is an risk of bias. The tool has been designed to comple-
appropriate period of washout between treatments. ment recent advancements in the field while main-
This may also be considered under question 6. taining its easy-to-follow questions. The revised JBI
Cluster RCTs randomize individuals or groups critical appraisal tool for RCTs offers systematic
(eg, communities, hospital wards), forming clusters. reviewers an improved and up-to-date method to
When we assess outcomes on an individual level in assess the risk of bias for RCTs included in their
cluster trials, there are unit-of-analysis issues, be- systematic review.
cause individuals within a cluster are correlated.
This should be considered by the study authors Acknowledgments
when conducting analysis, and ideally authors will
The JBI Scientific Committee members for their feed-
report the intra-cluster correlation coefficient. This
back and contributions regarding the concept of this
may also be considered under question 12.
work and both the draft and final manuscript.
Stepped-wedge RCTs may be appropriate to es-
Coauthor Catalin Tufanaru passed away July 29,
tablish when and how a beneficial intervention may
2021.
be best implemented within a defined setting, or due
to logistical, practical, or financial considerations in
the rollout of a new treatment/intervention. Data Funding
analysis in these trials should be conducted appro- MK is supported by the INTER-EXCELLENCE
priately, considering the effects of time. This may grant number LTC20031—Towards an Interna-
also be considered under question 12. tional Network for Evidence-based Research in Clin-
ical Health Research in the Czech Republic.
Conclusion ZM is supported by an NHMRC Investigator
Grant, APP1195676.
Randomized controlled studies are the ideal, and
often the only, included study design for systematic
reviews assessing the effectiveness of interventions.
References
All included studies must undergo rigorous critical
1. Munn Z, Barker TH, Moola S, Tufanaru C, Stern C, McArthur
appraisal, which, in the case of quantitative study A, et al. Methodological quality of case series studies: an
designs, is predominantly focused on assessment of introduction to the JBI critical appraisal tool. JBI Evid Synth
risk of bias in the conduct of the study. The revised 2020;18(10):2127–33.
JBI critical appraisal tool for RCTs presents an 2. Aromataris E, Munn Z. Chapter 1: JBI Systematic Reviews. In:
adaptable and robust new method for assessing this Aromataris E, Munn Z, editors. JBI Manual for Evidence
Synthesis [internet]. Adelaide, JBI; 2020 [cited 2022 Nov 29]. elaboration: updated guidance and exemplars for report-
Downloaded from http://journals.lww.com/jbisrir by BhDMf5ePHKav1zEoum1tQfN4a+kJLhEZgbsIHo4XMi0hCywCX
4. Tufanaru C, Munn Z, Aromataris E, Campbell J, Hopp L. Akl EA, Skoetz N, Guyatt GH. Chapter 14 Completing
Chapter 3: Systematic reviews of effectiveness. In: Aroma- ‘Summary of findings’ tables and grading the certainty
taris E, Munn Z, editors. JBI Manual for Evidence Synthesis of the evidence. In: Higgins JPT, Thomas J, Chandler J,
[internet]. Adelaide, JBI; 2020 [cited 2022 Nov 29]. Available Cumpston M, Li T, Page MJ, Welch VA, editors. Cochrane
from: https://synthesismanual.jbi.global. Handbook for Systematic Reviews of Interventions
5. Sterne JA, Savović J, Page MJ, Elbers RG, Blencowe NS, version 6.3 (updated February 2022). Cochrane, 2022
Boutron I, et al. RoB 2: a revised tool for assessing risk of [2022 Nov 29]. Available from: https://training.cochrane.
bias in randomised trials. BMJ 2019;366:1–8. org/handbook.
6. Stone JC, Glass K, Clark J, Munn Z, Tugwell P, Doi SAR. A 13. Aromataris E, Stern C, Lockwood C, Barker TH, Klugar M,
unified framework for bias assessment in clinical research. Jadotte Y, et al. JBI series paper 2: tailored evidence syn-
Int J Evid Based Healthc 2019;17(2):106–20. thesis approaches are required to answer diverse ques-
7. Stone JC, Gurunathan U, Aromataris E, Glass K, Tugwell P, tions: a pragmatic evidence synthesis toolkit from JBI.
Munn Z, et al. Bias assessment in outcomes research: the J Clin Epidemiol 2022;150:196–202.
role of relative versus absolute approaches. Value Health 14. Stone J, Gurunathan U, Glass K, Munn Z, Tugwell P, Doi
2021;24(8):1145–9. SAR. Stratification by quality induced selection bias in a
8. Barker TH, Stone JC, Sears K, Klugar M, Leonardi-Bee J, meta-analysis of clinical trials. J Clin Epidemiol 2019;107:
Tufanaru C, et al. Revising the JBI quantitative critical apprai- 51–9.
sal tools to improve their applicability: an overview of meth- 15. Greenland S, O'Rourke K. On the bias produced by quality
ods and the development process. JBI Evid Synth 2023;21(3): scores in meta-analysis, and a hierarchical view of pro-
478–93. posed solutions. Biostatistics 2001;2(4):463–71.
9. Jordan Z, Lockwood C, Aromataris E, Pilla B, Porritt K, Klugar 16. Stone JC, Glass K, Munn Z, Tugwell P, Doi SAR. Comparison
M, et al. JBI series paper 1: Introducing JBI and the JBI of bias adjustment methods in meta-analysis suggests
Model of EHBC. J Clin Epidemiol 2022;150:191–5. that quality effects modeling may have less limita
10. Page MJ, Moher D, Bossuyt PM, Boutron I, Hoffmann TC, tions than other approaches. J Clin Epidemiol 2020;117:
Mulrow CD, et al. PRISMA 2020 explanation and 36–45.