You are on page 1of 13

METHODOLOGY

The revised JBI critical appraisal tool for the assessment of


risk of bias for randomized controlled trials
Downloaded from http://journals.lww.com/jbisrir by BhDMf5ePHKav1zEoum1tQfN4a+kJLhEZgbsIHo4XMi0hCywCX

Timothy Hugh Barker1  Jennifer C. Stone1  Kim Sears2  Miloslav Klugar3  Catalin Tufanaru4  Jo Leonardi-Bee5
1AWnYQp/IlQrHD3i3D0OdRyi7TvSFl4Cf3VC1y0abggQZXdtwnfKZBYtws= on 11/04/2023

1 1
 Edoardo Aromataris  Zachary Munn

1
JBI, Faculty of Health and Medical Sciences, The University of Adelaide, Adelaide, SA, Australia, 2Queen’s Collaboration for Health Care Quality,
Queen’s University, Kingston, ON, Canada, 3Czech National Centre for Evidence-Based Healthcare and Knowledge Translation (Cochrane Czech
Republic, The Czech Republic (Middle European) Centre for Evidence-Based Healthcare: A JBI Centre of Excellence, Masaryk University GRADE
Centre), Faculty of Medicine, Institute of Biostatistics and Analyses, Masaryk University, Brno, Czech Republic, 4Centre for Health Informatics,
Australian Institute of Health Innovation, Macquarie University, Sydney, NSW, Australia, and 5The Nottingham Centre for Evidence Based
Healthcare: A JBI Centre of Excellence, School of Medicine, University of Nottingham, Nottingham, UK

ABSTRACT

JBI recently began the process of updating and revising its suite of critical appraisal tools to ensure that these tools
remain compatible with recent developments within risk of bias science. Following a rigorous development process
led by the JBI Effectiveness Methodology Group, this paper presents the revised critical appraisal tool for the
assessment of risk of bias for randomized controlled trials. This paper also presents practical guidance on how the
questions of this tool are to be interpreted and applied by systematic reviewers, while providing topical examples.
We also discuss the major changes made to this tool compared to the previous version and justification for why
these changes facilitate best-practice methodologies in this field.
Keywords: critical appraisal tool; methodological quality; methodology; randomized controlled trial; risk of bias
JBI Evid Synth 2023; 21(3):494–506.

Introduction appraisal.2,3 The purpose of this appraisal (for


quantitative evidence) is to determine the extent to
S ystematic reviews are a foundational and fun-
damental component in the practice of evidence-
based health care. Systematic reviews involve the
which a study has addressed the possibility of bias in
its design, conduct, or analysis. By subjecting every
collation and synthesis of the results of multiple in- study included in a systematic review to rigorous
dependent studies that address the same research critical appraisal, it allows reviewers to appropriately
question. Prior to the creation of these synthesized consider how the conduct of individual studies may
results, all studies that have been selected for inclu- impact the synthesized result, thus enabling the syn-
sion in the review (ie, those that meet the a priori thesized result to be correctly interpreted.4
eligibility criteria)1 must undergo a process of critical Recent advancements in the science of risk of bias
assessment5–7 have argued that only questions re-
lated to the internal validity of that study should
Correspondence: Timothy Hugh Barker, be considered in the assessment of that study’s
timothy.barker@adelaide.edu.au
Supplemental Digital Content is available for this article. Direct URL
inherent biases. The assessment of a study’s risk of
citations appear in the printed text and are provided in the HTML and bias often occurs during a structured and transpar-
PDF versions of this article on the journal's website, www.jbievi ent critical appraisal process. For example, a ques-
dencesynthesis.com. tion related to how generalizable a participant sam-
THB, JCS, EA, and ZM are paid employees of JBI, The University of ple is to the broader population does not impact on
Adelaide, and are members of the JBI Scientific Committee. THB, JCS, that study’s internal validity, and thus its inherent
KS, MK, JLB, EA, and ZM are members of the JBI Effectiveness Meth-
odology Group. MK is an associate editor, JLB is a senior associate
biases,5–8 but is still useful to describe the external
editor, and EA is editor in chief of JBI Evidence Synthesis. No authors validity of that study. There is also now an expecta-
were involved in the editorial processing of this manuscript. tion that assessments of bias occur at different levels,
DOI: 10.11124/JBIES-22-00430 including outcome level and result level assessments,

JBI Evidence Synthesis ß 2023 JBI 494

© 2023 JBI. Unauthorized reproduction of this article is prohibited.


METHODOLOGY T.H. Barker et al.

which may be different within the same study de- How to use the revised tool
Downloaded from http://journals.lww.com/jbisrir by BhDMf5ePHKav1zEoum1tQfN4a+kJLhEZgbsIHo4XMi0hCywCX

pending on the outcome or result being assessed.5,8 The key changes


These (and other) advancements have been dis- Similar to previous versions of these tools, the
cussed previously in an introduction to this body revised JBI critical appraisal tool for RCTs presents
of work.8
1AWnYQp/IlQrHD3i3D0OdRyi7TvSFl4Cf3VC1y0abggQZXdtwnfKZBYtws= on 11/04/2023

a series of questions. These questions aim to identify


It is acknowledged that the existing suite of JBI whether certain safeguards have been implemented
critical appraisal instruments are not aligned to these by the study to minimize risk of bias or to address
recent advancements, and conflate and confuse the other aspects relating to the validity or quality of the
process of critical appraisal with that of risk of bias study. Each question can be scored as being met
assessment. Therefore, the JBI Effectiveness Method- (yes), unmet (no), unclear, or not applicable. As
ology Group, under the auspices of the JBI Scientific described previously,8 the wording of the questions
Committee, updated the entire suite of JBI critical presented in the revised JBI critical appraisal tool for
appraisal tools to be better aligned to best-practice RCTs has not been altered from the wording of the
methodologies.8 This paper introduces the revised questions presented in the previous version of the JBI
critical appraisal tool for randomized controlled trials critical appraisal tool for RCTs.4 However, the
(RCTs) and provides step-by-step guidance on how to organization of these questions, the order in which
use and implement this tool in future systematic re- they should be addressed and answered, and the
views. We also clearly document and justify each means to answer them have been changed.
major change made in this revised tool. The questions of this revised tool have been
presented according to the construct of validity to
which they pertain. The specific validity constructs
Methods that are pertinent to the revised JBI critical appraisal
In 2021, a working group of researchers and metho- tool for RCTs include internal validity and statistical
dologists known as the JBI Effectiveness Method- conclusion validity. Questions that have been orga-
ology Group was tasked by the JBI Scientific Com- nized under the internal validity construct have been
mittee9 to revise the current suite of JBI critical further organized according to the domain of bias
appraisal tools for quantitative analytical study that they are specifically addressing. The domains of
designs. The aim of this work was to improve the bias relevant to the revised JBI critical appraisal tool
longevity and usefulness of these tools, and to for RCTs include bias related to selection and
reflect current advancements made in this space,5–7 allocation; administration of intervention/exposure;
while adhering to the reporting and methodological assessment, detection and measurement of the out-
requirements as established by PRISMA 202010 and come; and participant retention. A detailed descrip-
GRADE.11 To summarize this process, the JBI Effec- tion of these validity constructs and domains of
tiveness Methodology Group began by cataloguing biases is reported in a separate paper.8
the questions asked in each JBI critical appraisal tool The principal differences between the revised JBI
for study designs that employ quantitative data. critical appraisal tool for RCTs and its predecessor
These questions were ordered into constructs of are its structure and organization, which are now
validity (internal, statistical conclusion, comprehen- deliberately designed to facilitate judgments related
siveness of reporting, external) through a series of to risk of bias at different levels (eg, bias at the study
roundtable discussions between members of the JBI level, outcome level, or result level), where appropri-
Effectiveness Methodology Group. For questions that ate.8 For the questions that are to be answered at the
were related to the internal validity construct, they outcome level (questions 7–12), the tool provides the
were further catalogued to a domain of bias through ability to respond to the questions for up to 7 out-
a series of mapping exercises and roundtable discus- comes. The limit of 7 outcomes ensures that the tool
sions. Finally, questions were then separated based on aligns with the maximum number of outcomes
whether they were answered at the study, outcome, recommended to be included in a GRADE Summary
or result level. The full methodological processes of Findings or Evidence Profile.12 For the questions to
undertaken for this revision, including the rationale be answered at the result level (questions 10–12), the
for all decisions made, have been documented in a tool presents the option to record a different decision
separate paper.8 for 3 results for each outcome presented (by default).

JBI Evidence Synthesis ß 2023 JBI 495

© 2023 JBI. Unauthorized reproduction of this article is prohibited.


METHODOLOGY T.H. Barker et al.

Reviewers may face cases where there are fewer than recommend that studies be removed from a system-
Downloaded from http://journals.lww.com/jbisrir by BhDMf5ePHKav1zEoum1tQfN4a+kJLhEZgbsIHo4XMi0hCywCX

7 outcomes being appraised for a particular RCT, atic review following critical appraisal.
and there are more than 3 results being appraised per By removing studies, it presupposes that the pur-
outcome. The tool can be edited as required by the pose of a systematic review is to only permit the
1AWnYQp/IlQrHD3i3D0OdRyi7TvSFl4Cf3VC1y0abggQZXdtwnfKZBYtws= on 11/04/2023

review team to facilitate their use in these cases. synthesis of high-quality studies. While it may readily
For example, consider a hypothetical RCT that has promote alignment to the best available evidence, it
included 2 outcomes relevant to a question of a limits the full potential of the processes of evidence
systematic review team. These outcomes are mortal- synthesis to fully investigate eligible studies, their
ity and quality of life, both of which have been data, and provide a complete view of the evidence
measured at 2 time points within the study. When available to inform the review question.14,15 There are
using this tool, questions 1–6 and 13 are universal to several other approaches to incorporate the results
both outcomes, as they are addressed at the study of critical appraisal into the systematic review or
level and are only answered once. The reviewer meta-analysis. These approaches can include meta-
should then address question 7–9 twice, once for each regression, elicitation of expert opinion, using prior
outcome that is being appraised. Likewise, questions distributions, and quality-effect modeling.16 How-
10–12 should be addressed separately for both out- ever, these techniques demand appropriate statistical
comes but should also be assessed for each of the expertise and are beyond the scope of this paper.
results that has contributed data toward that out- Regardless of the approach ultimately chosen by the
come (eg, mortality at time point 1 and 2). In this reviewers, the results of the critical appraisal process
example, the reviewer would assess questions 10–12 should always be considered in the analysis and inter-
4 different times. It is also important to note that, as pretation of the findings of the synthesis.
with other critical appraisal tools,3,13 this tool should
also be applied in duplicate and independently during Overall assessment and presentation of results
the systematic review process. Reviewers should also Previous iterations of the JBI critical appraisal tool
be wary to only appraise outcomes that are relevant for RCTs intuitively supported reviewers assessing
to their systematic review question. If the only rele- the overall quality of a study using a checklist-based
vant outcome from this RCT for the systematic or scale-based tool structure (each item can be quan-
review question was mortality, then appraising the tified, which is enumerated to provide an overall
outcome quality of life would not be expected. quality score).8 The revised tool has been designed
to also facilitate judgments specific to the domains of
Interpretation of critical appraisal bias in which the questions belong
Some reviewers may take the approach of removing A reviewer may determine that for study 1, there
studies from progressing to data extraction or syn- was a low risk of bias for the domain “selection and
thesis in their review following the critical appraisal allocation,” as all questions received a response of
process. Removal of a study following critical ap- “yes.” However, for study 2, a reviewer may deter-
praisal may involve considering whether a certain mine a moderate risk of bias for the same domain, as
criterion had not been met (eg, randomization not the response to one of the questions was “no.” Im-
being demonstrated may warrant removal, assuming portantly, we provide no thresholds for grading of
the review was not also including other study de- bias severity (ie, low, moderate, high, critical, or other
signs with lesser internal validity due to not attempt- approaches) and leave this to the discretion of the
ing randomization). Another approach may include user and specific context in which they are working.
the review team weighting each question of the tool Considering the questions and assessments in this
(eg, randomization may be twice as important as regard, looking across all included studies (or a single
blinding of the outcome assessors); if a study fails study) can permit the reviewer to readily comment on
to meet a predetermined weight (decided by the how the risk of bias may impact on the certainty of
review team), then it may be removed. Other ap- their results at this domain level in the GRADE ap-
proaches may be to use simple cutoff scores (eg, if a proach. A judgment-based approach as described
study is scored with 10 “yes” responses, then it is here is one way for users to adopt the revised JBI
included) or to exclude studies that have been judged critical appraisal tool for RCTs; however, the tool is
as having a high risk of bias.8 However, we do not still compatible with either a checklist-based or

JBI Evidence Synthesis ß 2023 JBI 496

© 2023 JBI. Unauthorized reproduction of this article is prohibited.


METHODOLOGY T.H. Barker et al.

scale-based structure,8 and the decision of which ap- themselves. These known characteristics of the parti-
Downloaded from http://journals.lww.com/jbisrir by BhDMf5ePHKav1zEoum1tQfN4a+kJLhEZgbsIHo4XMi0hCywCX

proach to follow is left to the discretion of the cipants may distort the comparability of the groups
review team. (ie, does the intervention group contain more people
Current tools to appraise RCTs5 ask the reviewer over the age of 65 compared to the control?). A true
1AWnYQp/IlQrHD3i3D0OdRyi7TvSFl4Cf3VC1y0abggQZXdtwnfKZBYtws= on 11/04/2023

to establish an overall assessment of the risk of bias random assignment of participants to the groups
for each appraised study and for the overall body of means that a procedure is used that allocates the
evidence (ie, all appraised studies). The revised JBI participants to groups purely based on chance, not
critical appraisal tool for RCTs does not strictly influenced by any known characteristics of the parti-
prescribe to this, regardless of the approach fol- cipants. Reviewers should check the details about the
lowed. However, if reviewers opt to establish an randomization procedure used for allocation of the
overall assessment, then these assessments should participants to study groups. Was a true chance (ran-
not take into consideration the questions regarding dom) procedure used? For example, was a list of
statistical conclusion validity (questions 11–13) be- random numbers used? Was a computer-generated
cause risk of bias is only impacted by the internal list of random numbers used? Was a statistician,
validity construct.8 external to the research team, consulted for the
Irrespective of the approach taken, the results of randomization sequence generation? Additionally,
critical appraisal using the revised JBI critical apprai- reviewers should check that the authors are not stat-
sal tool for RCTs should be reported narratively in ing they have used random approaches when they
the review. This narrative summary should provide have instead used systematic approaches (such as
both an overall description of the methodological allocating by days of the week).
quality and risk of bias at the domain level of the
Question 2: Was allocation to groups concealed?
included studies. There should also be a statement
Category: Internal validity
regarding any important or interesting deviations
Domain: Bias related to selection and allocation
from the observed trends. This narrative summary
Appraisal: Study level
can be supported with the use of a table or graphic
If those allocating participants to the compared
that shows how each included study responded. We
groups are aware of which group is next in the
recommend presenting the results of critical appraisal
allocation process (ie, the treatment or control
for all questions via a table rather than summarizing
group), there is a risk that they may deliberately
with a score; see the example in Table 1. (Note that
and purposefully intervene in the allocation of pa-
this design is not prescriptive and only serves as an
tients. This may result in the preferential allocation
example).
of patients to the treatment group or to the control
group. This may directly distort the results of the
The revised JBI critical appraisal tool for study, as participants no longer have an equal and
randomized controlled trials random chance to belong to each group. Conceal-
The criteria and considerations that should be made ment of allocation refers to procedures that prevent
by reviewers when answering the questions in the those allocating patients from knowing before allo-
revised JBI critical appraisal tool for RCTs are cation which treatment or control is next in the
shown in Table 2. This tool is also available to allocation process. Reviewers should check the de-
download as Supplemental Digital Content 1 at tails about the procedure used for allocation con-
http://links.lww.com/SRX/A7. cealment. Was an appropriate allocation conceal-
ment procedure used? For example, was central
Question 1: Was true randomization used for assign-
randomization used? Were sequentially numbered,
ment of participants to treatment groups?
opaque, and sealed envelopes used? Were coded
Category: Internal validity
drug packs used?
Domain: Bias related to selection and allocation
Appraisal: Study level Question 3: Were treatment groups similar at the
If participants are not allocated to treatment and baseline?
control groups by random assignment, there is a risk Category: Internal validity
that this assignment to groups can be influenced by Domain: Bias related to selection and allocation
the known characteristics of the participants Appraisal: Study level

JBI Evidence Synthesis ß 2023 JBI 497

© 2023 JBI. Unauthorized reproduction of this article is prohibited.


1AWnYQp/IlQrHD3i3D0OdRyi7TvSFl4Cf3VC1y0abggQZXdtwnfKZBYtws= on 11/04/2023
Downloaded from http://journals.lww.com/jbisrir by BhDMf5ePHKav1zEoum1tQfN4a+kJLhEZgbsIHo4XMi0hCywCX

JBI Evidence Synthesis

METHODOLOGY
© 2023 JBI. Unauthorized reproduction of this article is prohibited.

Table 1: Example presentation of results following critical appraisal using the revised JBI critical appraisal tool for randomized
controlled trials

INTERNAL VALIDITY Bias related to:

Assessment, detection, and Partici-


Administration of measurement of the pant STATISTICAL CONCLUSION
DOMAIN Selection and allocation intervention/exposure outcome retention VALIDITY

QUESTION NO. 1 2 3 4 5 6 7 8 9 10 11 12 13

STUDY ID OUTCOME RESULT

Study 1 Mortality Time 1 Y Y Y N Y Y Y Y Y Y Y Y Y


Mortality Time 2 Y Y Y
QOL Time 1 N Y Y Y Y Y
QOL Time 2 N Y Y
Study 2 Mortality Time 1 Y Y N Y Y Y Y Y Y Y Y Y Y
Mortality Time 2 Y Y Y
QOL Time 1 Y Y Y Y Y Y
QOL Time 2 Y Y Y
QOL, quality of life
Example of how the results of critical appraisal may be presented when using the revised JBI critical appraisal tool for randomized controlled trials. This example presents the results that clearly distinguish the relationship of
the result to the outcome, and the outcome to the study. Reviewers can also provide summary judgments for each domain of bias and validity construct presented. For example, for study 1, there may be a low risk of bias for
the domain of selection and allocation, as all questions responded with “yes.” However, study 2 may be considered to have moderate risk of bias for the same domain, as the response to one of the questions was “no.”
ß 2023 JBI

T.H. Barker et al.


498
METHODOLOGY T.H. Barker et al.

As with question 1, any difference between the Appraisal: Study level


Downloaded from http://journals.lww.com/jbisrir by BhDMf5ePHKav1zEoum1tQfN4a+kJLhEZgbsIHo4XMi0hCywCX

known characteristics of participants included in the Like question 4, those delivering the treatment who
compared groups constitutes a threat to internal va- are aware of participant allocation to either treat-
lidity. If differences in these characteristics do ment or control may treat participants differently
1AWnYQp/IlQrHD3i3D0OdRyi7TvSFl4Cf3VC1y0abggQZXdtwnfKZBYtws= on 11/04/2023

exist, then there is potential that the effect cannot compared to those who remain unaware of parti-
be attributed to the potential cause (the examined cipant allocation. There is a risk that any potential
intervention or treatment). This is because the effect change in behavior may influence the implementa-
may be explained by the differences between partici- tion of the compared treatments, and the results of
pant characteristics and not the intervention/treat- the study may be distorted. Blinding of those deli-
ment of interest. Reviewers should check the charac- vering treatment is used to minimize this risk.
teristics reported for participants. Are the participants When this level of blinding has been achieved,
from the compared groups similar with regard to the those delivering the treatment are not aware if they
characteristics that may explain the effect, even in the are treating the group receiving the treatment of
absence of the cause (eg, age, severity of the disease, interest or if they are treating any other group
stage of the disease, coexisting conditions)? Re- receiving the control intervention. Reviewers
viewers should check the proportion of participants should check the details reported in the article
with specific relevant characteristics in the compared about the blinding of those delivering treatment
groups. (Note: Do not only consider the P value for with regard to treatment assignment. Is there any
the statistical testing of the differences between
information in the article about those delivering
groups with regard to the baseline characteristics.)
the treatment? Were those delivering the treatment
Question 4: Were participants blind to treatment unaware of the assignments of participants to the
assignment? compared groups?
Category: Internal validity
Question 6: Were treatment groups treated identi-
Domain: Bias related to administration of interven-
cally other than the intervention of interest?
tion/exposure
Category: Internal validity
Appraisal: Study level
Domain: Bias related to administration of interven-
Participants who are aware of their allocation to
tion/exposure
either the treatment or the control may behave, re-
Appraisal: Study level
spond, or react differently to their assigned treatment
To attribute the effect to the cause (assuming there
(or control) compared with participants who remain
is no bias related to selection and allocation), there
unaware of their allocation. Blinding of participants
should be no difference between the groups in
is a technique used to minimize this risk. Blinding
terms of treatment or care received, other than
refers to procedures that prevent participants from
the treatment or intervention controlled by the
knowing which group they are allocated. If blinding
researchers. If there are other exposures or treat-
has been followed, participants are not aware if they
ments occurring at the same time as the cause (the
are in the group receiving the treatment of interest or
treatment or intervention of interest), then the ef-
if they are in another group receiving the control
fect can potentially be attributed to something
intervention. Reviewers should check the details re-
other than the examined cause (the investigated
ported in the article about the blinding of participants
treatment). This is because it is plausible that the
with regard to treatment assignment. Was an appro-
effect may be explained by other exposures or
priate blinding procedure used? For example, were
treatments that occurred at the same time as the
identical capsules or syringes used? Were identical
cause. Reviewers should check the reported expo-
devices used? Be aware of different terms used; blind-
sures or interventions received by the compared
ing is sometimes also called masking.
groups. Are there other exposures or treatments
Question 5: Were those delivering the treatment occurring at the same time as the cause? Is it
blind to treatment assignment? plausible that the effect may be explained by other
Category: Internal validity exposures or treatments occurring at the same time
Domain: Bias related to administration of interven- as the cause? Is it clear that there is no other
tion/exposure difference between the groups in terms of treatment

JBI Evidence Synthesis ß 2023 JBI 499

© 2023 JBI. Unauthorized reproduction of this article is prohibited.


METHODOLOGY T.H. Barker et al.

or care received, other than the treatment or inter- statistical relationship between the cause and the
Downloaded from http://journals.lww.com/jbisrir by BhDMf5ePHKav1zEoum1tQfN4a+kJLhEZgbsIHo4XMi0hCywCX

vention of interest? effect estimated in a study exploring causal effects.


Unreliability of outcome measurements is one of the
Question 7: Were outcome assessors blind to treat- plausible explanations for errors of statistical infer-
ment assignment?
1AWnYQp/IlQrHD3i3D0OdRyi7TvSFl4Cf3VC1y0abggQZXdtwnfKZBYtws= on 11/04/2023

ence with regard to the existence and the magnitude


Category: Internal validity of the effect determined by the treatment (cause).
Domain: Bias related to assessment, detection, and Reviewers should check the details about the relia-
measurement of the outcome bility of the measurement used, such as the number
Appraisal: Outcome level of raters, the training of raters, and the reliability of
Like questions 4 and 5, if those assessing the out- the intra-rater and the inter-raters within the study
comes are aware of participant allocation to either (not as reported in external sources). This question is
treatment or control, they may treat participants about the reliability of the measurement performed
differently compared with those who remain una- in the study, and not about the validity of the mea-
ware of participant allocation. Therefore, there is a surement instruments/scales used in the study. Fi-
risk that the measurement of the outcomes between nally, some outcomes may not rely on instruments
groups may be distorted, and the results of the study or scales (eg, death), and reliability of the measure-
may themselves be distorted. Blinding of outcomes ments may need to be assessed in the context of the
assessors is used to minimize this risk. Reviewers study being reviewed. (Note: Two other important
should check the details reported in the article about threats that weaken the validity of inferences about
the blinding of outcomes assessors with regard to the statistical relationship between the cause and the
treatment assignment. Is there any information in effect are low statistical power and the violation of
the article about outcomes assessors? Were those the assumptions of statistical tests. These threats are
assessing the treatment’s effects on outcomes una- explored within question 12.)
ware of the assignments of participants to the com-
Question 10: Was follow-up complete and, if not,
pared groups?
were differences between groups in terms of their
Question 8: Were outcomes measured in the same follow-up adequately described and analyzed?
way for treatment groups? Category: Internal validity
Category: Internal validity Domain: Bias related to participant retention
Domain: Bias related to assessment, detection, and Appraisal: Result level
measurement of the outcome For this question, follow-up refers to the period from
Appraisal: Outcome level the moment of randomization to any point in which
If the outcome is not measured in the same way in the the groups are compared during the trial. This ques-
compared groups, there is a threat to the internal tion asks whether there is complete knowledge (eg,
validity of a study. Any differences in outcome mea- measurements, observations) for the entire duration
surements may be due to the method of measurement of the trial for all randomly allocated participants. If
employed between the 2 groups and not the interven- there is incomplete follow-up from all randomly
tion/treatment of interest. Reviewers should check allocated participants, this is known as post-assign-
whether the outcomes were measured in the same ment attrition. Because RCTs are not perfect, there is
way. Was the same instrument or scale used? Was almost always post-assignment attrition, and the
the measurement timing the same? Were the measure- focus of this question is on the appropriate explora-
ment procedures and instructions the same? tion of post-assignment attrition. If differences exist
Question 9: Were outcomes measured in a reliable with regard to the post-assignment attrition between
way? the compared groups of an RCT, then there is a
Category: Internal validity threat to the internal validity of that study. This is
Domain: Bias related to assessment, detection, and because these differences may provide a plausible
measurement of the outcome alternative explanation for the observed effect even
Appraisal: Outcome level in the absence of the cause (the treatment or inter-
Unreliability of outcome measurements is one threat vention of interest). It is important to note that with
that weakens the validity of inferences about the regard to post-assignment attrition, it is not enough

JBI Evidence Synthesis ß 2023 JBI 500

© 2023 JBI. Unauthorized reproduction of this article is prohibited.


METHODOLOGY T.H. Barker et al.

to know the number of participants and the propor- Were participants analyzed in the groups to which
Downloaded from http://journals.lww.com/jbisrir by BhDMf5ePHKav1zEoum1tQfN4a+kJLhEZgbsIHo4XMi0hCywCX

tions of participants with incomplete data; the rea- they were initially randomized, regardless of
sons for loss to follow-up are essential in the analysis whether they participated in those groups and
of risk of bias. regardless of whether they received the planned
1AWnYQp/IlQrHD3i3D0OdRyi7TvSFl4Cf3VC1y0abggQZXdtwnfKZBYtws= on 11/04/2023

Reviewers should check whether there were dif- interventions?


ferences with regard to the loss to follow-up between Note: The ITT analysis is a type of statistical
the compared groups. If follow-up was incomplete analysis recommended in the Consolidated Stan-
(incomplete information on all participants), exam- dards of Reporting Trials (CONSORT) statement
ine the reported details about the strategies used to on best practices in trials reporting, and it is consid-
address incomplete follow-up. This can include de- ered a marker of good methodological quality of the
scriptions of loss to follow-up (eg, absolute numbers, analysis of results of a randomized trial. The ITT is
proportions, reasons for loss to follow-up) and im- estimating the effect of offering the intervention (ie,
pact analyses (the analyses of the impact of loss to the effect of instructing the participants to use or
follow-up on results). Was there a description of the take the intervention); the ITT is not estimating the
incomplete follow-up including the number of parti- effect of receiving the intervention of interest.
cipants and the specific reasons for loss to follow-
up? Even if follow-up was incomplete but balanced Question 12: Was appropriate statistical analysis
between groups, if the reasons for loss to follow-up used?
are different (eg, side effects caused by the interven- Category: Statistical conclusion validity
tion of interest), these may impose a risk of bias if Appraisal: Result level
not appropriately explored in the analysis. If there Inappropriate statistical analysis may cause errors of
are differences between groups with regard to the statistical inference with regard to the existence and
loss to follow-up (numbers/proportions and rea- the magnitude of the effect determined by the treat-
sons), was there an analysis of patterns of loss to ment (cause). Low statistical power and the violation
follow-up? If there are differences between the groups of the assumptions of statistical tests are 2 important
with regard to the loss to follow-up, was there an threats that weaken the validity of inferences about
analysis of the impact of the loss to follow-up on the statistical relationship between the cause and the
the results? (Note: Question 10 is not about inten- effect. Reviewers should check the following aspects:
tion-to-treat [ITT] analysis; question 11 is about ITT if the assumptions of the statistical tests were re-
analysis.) spected; if appropriate statistical power analysis
was performed; if appropriate effect sizes were used;
Question 11: Were participants analyzed in the if appropriate statistical methods were used given
groups to which they were randomized? the nature of the data and the objectives of statistical
Category: Statistical conclusion validity analysis (eg, association between variables, predic-
Appraisal: Result level tion, survival analysis).
This question is about the ITT analysis. There are
different statistical analysis strategies available for Question 13: Was the trial design appropriate and
the analysis of data from RCTs, such as ITT, per- any deviations from the standard RCT design (indi-
protocol analysis, and as-treated analysis. In the ITT vidual randomization, parallel groups) accounted
analysis, the participants are analyzed in the groups for in the conduct and analysis of the trial?
to which they were randomized. This means that Category: Statistical conclusion validity
regardless of whether participants received the Appraisal: Study level
intervention or control as assigned, were compliant The typical, parallel group RCT may not always be
with their planned assignment, or participated for appropriate depending on the nature of the question.
the entire study duration, they are still included in Therefore, some additional RCT designs may have
the analysis. The ITT analysis compares the out- been employed that come with their own additional
comes for participants from the initial groups cre- considerations.
ated by the initial random allocation of participants Crossover trials should only be conducted with
to those groups. Reviewers should check whether an people with a chronic, stable condition, where the
ITT analysis was reported and the details of the ITT. intervention produces a short-term effect (eg, relief in

JBI Evidence Synthesis ß 2023 JBI 501

© 2023 JBI. Unauthorized reproduction of this article is prohibited.


METHODOLOGY T.H. Barker et al.

Table 2: JBI critical appraisal tool for randomized controlled trials


Downloaded from http://journals.lww.com/jbisrir by BhDMf5ePHKav1zEoum1tQfN4a+kJLhEZgbsIHo4XMi0hCywCX

RoB assessor: Date of appraisal: Record number:

Study author: Study title: Study year:


1AWnYQp/IlQrHD3i3D0OdRyi7TvSFl4Cf3VC1y0abggQZXdtwnfKZBYtws= on 11/04/2023

Choice - comments/
Internal validity justification Yes No Unclear N/A

Bias related to selection and allocation


1 Was true randomization used for assignment □ □ □ □
of participants to treatment groups?
2 Was allocation to treatment groups □ □ □ □
concealed?
3 Were treatment groups similar at the □ □ □ □
baseline?
Bias related to administration of intervention/exposure
4 Were participants blind to treatment □ □ □ □
assignment?
5 Were those delivering the treatment blind to □ □ □ □
treatment assignment?
6 Were treatment groups treated identically □ □ □ □
other than the intervention of interest?
Bias related to assessment, detection, and measurement of the outcome
7 Were outcome assessors blind to treatment Yes No Unclear N/A
assignment?
Outcome 1 □ □ □ □
Outcome 2 □ □ □ □
Outcome 3 □ □ □ □
Outcome 4 □ □ □ □
Outcome 5 □ □ □ □
Outcome 6 □ □ □ □
Outcome 7 □ □ □ □
8 Were outcomes measured in the same way for Yes No Unclear N/A
treatment groups?
Outcome 1 □ □ □ □
Outcome 2 □ □ □ □
Outcome 3 □ □ □ □
Outcome 4 □ □ □ □
Outcome 5 □ □ □ □
Outcome 6 □ □ □ □
Outcome 7 □ □ □ □
9 Were outcomes measured in a reliable way? Yes No Unclear N/A
Outcome 1 □ □ □ □
Outcome 2 □ □ □ □
Outcome 3 □ □ □ □
Outcome 4 □ □ □ □
Outcome 5 □ □ □ □
Outcome 6 □ □ □ □
Outcome 7 □ □ □ □

JBI Evidence Synthesis ß 2023 JBI 502

© 2023 JBI. Unauthorized reproduction of this article is prohibited.


METHODOLOGY T.H. Barker et al.

Table 2: (continued )
Downloaded from http://journals.lww.com/jbisrir by BhDMf5ePHKav1zEoum1tQfN4a+kJLhEZgbsIHo4XMi0hCywCX

Choice - comments/
Internal validity justification Yes No Unclear N/A
1AWnYQp/IlQrHD3i3D0OdRyi7TvSFl4Cf3VC1y0abggQZXdtwnfKZBYtws= on 11/04/2023

Bias related to participant retention


10 Was follow-up complete and, if not, were differences between groups in terms of their follow-up adequately described and analyzed?
Outcome 1 Yes No Unclear N/A
Result 1 □ □ □ □
Result 2 □ □ □ □
Result 3 □ □ □ □
Outcome 2 Yes No Unclear N/A
Result 1 □ □ □ □
Result 2 □ □ □ □
Result 3 □ □ □ □
Outcome 3 Yes No Unclear N/A
Result 1 □ □ □ □
Result 2 □ □ □ □
Result 3 □ □ □ □
Outcome 4 Yes No Unclear N/A
Result 1 □ □ □ □
Result 2 □ □ □ □
Result 3 □ □ □ □
Outcome 5 Yes No Unclear N/A
Result 1 □ □ □ □
Result 2 □ □ □ □
Result 3 □ □ □ □
Outcome 6 Yes No Unclear N/A
Result 1 □ □ □ □
Result 2 □ □ □ □
Result 3 □ □ □ □
Outcome 7 Yes No Unclear N/A
Result 1 □ □ □ □
Result 2 □ □ □ □
Result 3 □ □ □ □

Choice - comments/
Statistical conclusion validity justification Yes No Unclear N/A

11 Were participants analyzed in the groups to which they were randomized?


Outcome 1 Yes No Unclear N/A
Result 1 □ □ □ □
Result 2 □ □ □ □
Result 3 □ □ □ □
Outcome 2 Yes No Unclear N/A
Result 1 □ □ □ □
Result 2 □ □ □ □
Result 3 □ □ □ □
Outcome 3 Yes No Unclear N/A
Result 1 □ □ □ □

JBI Evidence Synthesis ß 2023 JBI 503

© 2023 JBI. Unauthorized reproduction of this article is prohibited.


METHODOLOGY T.H. Barker et al.

Table 2: (continued )
Downloaded from http://journals.lww.com/jbisrir by BhDMf5ePHKav1zEoum1tQfN4a+kJLhEZgbsIHo4XMi0hCywCX

Choice - comments/
Statistical conclusion validity justification Yes No Unclear N/A
1AWnYQp/IlQrHD3i3D0OdRyi7TvSFl4Cf3VC1y0abggQZXdtwnfKZBYtws= on 11/04/2023

Result 2 □ □ □ □
Result 3 □ □ □ □
Outcome 4 Yes No Unclear N/A
Result 1 □ □ □ □
Result 2 □ □ □ □
Result 3 □ □ □ □
Outcome 5 Yes No Unclear N/A
Result 1 □ □ □ □
Result 2 □ □ □ □
Result 3 □ □ □ □
Outcome 6 Yes No Unclear N/A
Result 1 □ □ □ □
Result 2 □ □ □ □
Result 3 □ □ □ □
Outcome 7 Yes No Unclear N/A
Result 1 □ □ □ □
Result 2 □ □ □ □
Result 3 □ □ □ □
12 Was appropriate statistical analysis used?
Outcome 1 Yes No Unclear N/A
Result 1 □ □ □ □
Result 2 □ □ □ □
Result 3 □ □ □ □
Outcome 2 Yes No Unclear N/A
Result 1 □ □ □ □
Result 2 □ □ □ □
Result 3 □ □ □ □
Outcome 3 Yes No Unclear N/A
Result 1 □ □ □ □
Result 2 □ □ □ □
Result 3 □ □ □ □
Outcome 4 Yes No Unclear N/A
Result 1 □ □ □ □
Result 2 □ □ □ □
Result 3 □ □ □ □
Outcome 5 Yes No Unclear N/A
Result 1 □ □ □ □
Result 2 □ □ □ □
Result 3 □ □ □ □
Outcome 6 Yes No Unclear N/A
Result 1 □ □ □ □
Result 2 □ □ □ □
Result 3 □ □ □ □

JBI Evidence Synthesis ß 2023 JBI 504

© 2023 JBI. Unauthorized reproduction of this article is prohibited.


METHODOLOGY T.H. Barker et al.

Table 2: (continued )
Downloaded from http://journals.lww.com/jbisrir by BhDMf5ePHKav1zEoum1tQfN4a+kJLhEZgbsIHo4XMi0hCywCX

Choice - comments/
Statistical conclusion validity justification Yes No Unclear N/A
1AWnYQp/IlQrHD3i3D0OdRyi7TvSFl4Cf3VC1y0abggQZXdtwnfKZBYtws= on 11/04/2023

Outcome 7 Yes No Unclear N/A


Result 1 □ □ □ □
Result 2 □ □ □ □
Result 3 □ □ □ □
13 Was the trial design appropriate and any □ □ □ □
deviations from the standard RCT design
(individual randomization, parallel groups)
accounted for in the conduct and analysis of
the trial?
Overall Include: □ Exclude: □ Seek further info: □
appraisal:
Comments:

symptoms). Crossover trials should ensure there is an risk of bias. The tool has been designed to comple-
appropriate period of washout between treatments. ment recent advancements in the field while main-
This may also be considered under question 6. taining its easy-to-follow questions. The revised JBI
Cluster RCTs randomize individuals or groups critical appraisal tool for RCTs offers systematic
(eg, communities, hospital wards), forming clusters. reviewers an improved and up-to-date method to
When we assess outcomes on an individual level in assess the risk of bias for RCTs included in their
cluster trials, there are unit-of-analysis issues, be- systematic review.
cause individuals within a cluster are correlated.
This should be considered by the study authors Acknowledgments
when conducting analysis, and ideally authors will
The JBI Scientific Committee members for their feed-
report the intra-cluster correlation coefficient. This
back and contributions regarding the concept of this
may also be considered under question 12.
work and both the draft and final manuscript.
Stepped-wedge RCTs may be appropriate to es-
Coauthor Catalin Tufanaru passed away July 29,
tablish when and how a beneficial intervention may
2021.
be best implemented within a defined setting, or due
to logistical, practical, or financial considerations in
the rollout of a new treatment/intervention. Data Funding
analysis in these trials should be conducted appro- MK is supported by the INTER-EXCELLENCE
priately, considering the effects of time. This may grant number LTC20031—Towards an Interna-
also be considered under question 12. tional Network for Evidence-based Research in Clin-
ical Health Research in the Czech Republic.
Conclusion ZM is supported by an NHMRC Investigator
Grant, APP1195676.
Randomized controlled studies are the ideal, and
often the only, included study design for systematic
reviews assessing the effectiveness of interventions.
References
All included studies must undergo rigorous critical
1. Munn Z, Barker TH, Moola S, Tufanaru C, Stern C, McArthur
appraisal, which, in the case of quantitative study A, et al. Methodological quality of case series studies: an
designs, is predominantly focused on assessment of introduction to the JBI critical appraisal tool. JBI Evid Synth
risk of bias in the conduct of the study. The revised 2020;18(10):2127–33.
JBI critical appraisal tool for RCTs presents an 2. Aromataris E, Munn Z. Chapter 1: JBI Systematic Reviews. In:
adaptable and robust new method for assessing this Aromataris E, Munn Z, editors. JBI Manual for Evidence

JBI Evidence Synthesis ß 2023 JBI 505

© 2023 JBI. Unauthorized reproduction of this article is prohibited.


METHODOLOGY T.H. Barker et al.

Synthesis [internet]. Adelaide, JBI; 2020 [cited 2022 Nov 29]. elaboration: updated guidance and exemplars for report-
Downloaded from http://journals.lww.com/jbisrir by BhDMf5ePHKav1zEoum1tQfN4a+kJLhEZgbsIHo4XMi0hCywCX

Available from: https://synthesismanual.jbi.global. ing systematic reviews. BMJ 2021;372:n160.


3. Porritt K, Gomersall J, Lockwood C. JBI's systematic reviews: 11. GRADE Working Group. Grading quality of evidence and
study selection and critical appraisal. Am J Nurs 2014;114 strength of recommendations. BMJ 2004;328(7454):1490.
(6):47–52. 12. Schünemann HJ, Higgins JP, Vist GE, Glasziou P,
1AWnYQp/IlQrHD3i3D0OdRyi7TvSFl4Cf3VC1y0abggQZXdtwnfKZBYtws= on 11/04/2023

4. Tufanaru C, Munn Z, Aromataris E, Campbell J, Hopp L. Akl EA, Skoetz N, Guyatt GH. Chapter 14 Completing
Chapter 3: Systematic reviews of effectiveness. In: Aroma- ‘Summary of findings’ tables and grading the certainty
taris E, Munn Z, editors. JBI Manual for Evidence Synthesis of the evidence. In: Higgins JPT, Thomas J, Chandler J,
[internet]. Adelaide, JBI; 2020 [cited 2022 Nov 29]. Available Cumpston M, Li T, Page MJ, Welch VA, editors. Cochrane
from: https://synthesismanual.jbi.global. Handbook for Systematic Reviews of Interventions
5. Sterne JA, Savović J, Page MJ, Elbers RG, Blencowe NS, version 6.3 (updated February 2022). Cochrane, 2022
Boutron I, et al. RoB 2: a revised tool for assessing risk of [2022 Nov 29]. Available from: https://training.cochrane.
bias in randomised trials. BMJ 2019;366:1–8. org/handbook.
6. Stone JC, Glass K, Clark J, Munn Z, Tugwell P, Doi SAR. A 13. Aromataris E, Stern C, Lockwood C, Barker TH, Klugar M,
unified framework for bias assessment in clinical research. Jadotte Y, et al. JBI series paper 2: tailored evidence syn-
Int J Evid Based Healthc 2019;17(2):106–20. thesis approaches are required to answer diverse ques-
7. Stone JC, Gurunathan U, Aromataris E, Glass K, Tugwell P, tions: a pragmatic evidence synthesis toolkit from JBI.
Munn Z, et al. Bias assessment in outcomes research: the J Clin Epidemiol 2022;150:196–202.
role of relative versus absolute approaches. Value Health 14. Stone J, Gurunathan U, Glass K, Munn Z, Tugwell P, Doi
2021;24(8):1145–9. SAR. Stratification by quality induced selection bias in a
8. Barker TH, Stone JC, Sears K, Klugar M, Leonardi-Bee J, meta-analysis of clinical trials. J Clin Epidemiol 2019;107:
Tufanaru C, et al. Revising the JBI quantitative critical apprai- 51–9.
sal tools to improve their applicability: an overview of meth- 15. Greenland S, O'Rourke K. On the bias produced by quality
ods and the development process. JBI Evid Synth 2023;21(3): scores in meta-analysis, and a hierarchical view of pro-
478–93. posed solutions. Biostatistics 2001;2(4):463–71.
9. Jordan Z, Lockwood C, Aromataris E, Pilla B, Porritt K, Klugar 16. Stone JC, Glass K, Munn Z, Tugwell P, Doi SAR. Comparison
M, et al. JBI series paper 1: Introducing JBI and the JBI of bias adjustment methods in meta-analysis suggests
Model of EHBC. J Clin Epidemiol 2022;150:191–5. that quality effects modeling may have less limita
10. Page MJ, Moher D, Bossuyt PM, Boutron I, Hoffmann TC, tions than other approaches. J Clin Epidemiol 2020;117:
Mulrow CD, et al. PRISMA 2020 explanation and 36–45.

JBI Evidence Synthesis ß 2023 JBI 506

© 2023 JBI. Unauthorized reproduction of this article is prohibited.

You might also like