You are on page 1of 9

Journal of Evidence-Based Medicine ISSN 1756-5391

REVIEW ARTICLE

The methodological quality assessment tools for preclinical


and clinical studies, systematic review and meta-analysis,
and clinical practice guideline: a systematic review
Xiantao Zeng1,2 , Yonggang Zhang3 , Joey S.W. Kwong3 , Chao Zhang2 , Sheng Li1 , Feng Sun4 ,
Yuming Niu2 and Liang Du3
1
Center for Evidence-Based and Translational Medicine, Zhongnan Hospital, Wuhan University, Wuhan, China
2
Center for Evidence-based Medicine and Clinical Research, Taihe Hospital, Hubei University of Medicine, Shiyan, China
3
Chinese Cochrane Centre, West China Hospital, Sichuan University, Chengdu, China
4
Department of Epidemiology and Biostatistics, School of Public Health, Peking University Health Science Centre, Beijing, China

Keywords Abstract
Clinical practice guideline; meta-analysis;
methodological quality; primary study; risk of Objective: To systematically review the methodological assessment tools for pre-
bias; systematic review. clinical and clinical studies, systematic review and meta-analysis, and clinical
practice guideline.
Correspondence
Liang Du, Chinese Cochrane Centre, West
Methods: We searched PubMed, the Cochrane Handbook for Systematic Reviews
China Hospital, Sichuan University, of Interventions, Joanna Briggs Institute (JBI) Reviewers Manual, Centre for Re-
Guoxuexiang 37#, Chengdu 610041, China. views and Dissemination, Critical Appraisal Skills Programme (CASP), Scottish
Tel: 86-28-85426295; Intercollegiate Guidelines Network (SIGN), and the National Institute for Clinical
Fax: 86-28-85426295; Excellence (NICE) up to May 20th, 2014. Two authors selected studies and ex-
Email: duliang0606@vip.sina.com tracted data; quantitative analysis was performed to summarize the characteristics
of included tools.
Conflict of interest
Results: We included a total of 21 assessment tools for analysis. A number of tools
The authors declare no conflict of interest.
were developed by academic organizations, and some were developed by only a
Received 1 December 2014; accepted for small group of researchers. The JBI developed the highest number of method-
publication 10 December 2014. ological assessment tools, with CASP coming second. Tools for assessing the
methodological quality of randomized controlled studies were most abundant. The
doi: 10.1111/jebm.12141 Cochrane Collaboration’s tool for assessing risk of bias is the best available tool
for assessing RCTs. For cohort and case-control studies, we recommend the use
of the Newcastle-Ottawa Scale. The Methodological Index for Non-Randomized
Studies (MINORS) is an excellent tool for assessing non-randomized interventional
studies, and the Agency for Healthcare Research and Quality (ARHQ) methodol-
ogy checklist is applicable for cross-sectional studies. For diagnostic accuracy test
studies, the Quality Assessment of Diagnostic Accuracy Studies-2 (QUADAS-2)
tool is recommended; the SYstematic Review Centre for Laboratory animal Exper-
imentation (SYRCLE) risk of bias tool is available for assessing animal studies;
Assessment of Multiple Systematic Reviews (AMSTAR) is a measurement tool for
systematic reviews/meta-analyses; an 18-item tool has been developed for apprais-
ing case series studies, and the Appraisal of Guidelines, Research and Evaluation
(AGREE)-II instrument is widely used to evaluate clinical practice guidelines.
Conclusions: We have successfully identified a variety of methodological assess-
ment tools for different types of study design. However, further efforts in the de-
velopment of critical appraisal tools are warranted since there is currently a lack of
such tools for other fields, e.g. genetic studies, and some existing tools (nested case-
control studies and case reports, for example) are in need of updating to be in line
with current research practice and rigor. In addition, it is very important that all crit-
ical appraisal tools remain subjective and performance bias is effectively avoided.

2 JEBM 8 (2015) 2–10 


C 2015 Chinese Cochrane Center, West China Hospital of Sichuan University and Wiley Publishing Asia Pty Ltd
X. Zeng et al. Methodological quality assessment tools

Figure 1 An overview of the different study types in medical research.

Introduction and other biases during the research process (5). Therefore,
all methodological quality assessment tools are focused on
Professor Archibald (Archie) Cochrane (1909 to 1988) these aspects, which is recommended as “risk of bias” by the
is considered to be the pioneer and father of the idea Cochrane Handbook for Systematic Reviews of Interventions
of evidence-based medicine (EBM) in our era (1). The (the Cochrane Handbook) (5). There are three types of tools:
concept of “EBM” is first formally put forward in 1992 scales, checklists, and items (Table 1) (8). Correctly choosing
by the Evidence-Based Medicine Working Group, which the applicability of the tools is very important; therefore, the
is led by Professor Gordon H. Guyatt (2). One year later, purpose of our study was to systematically review the features
Professor Ian Chalmers, the student of Archie Cochrane and usability of the currently available and commonly used
and a group of 70 other international colleagues found The methodological quality assessment tools for preclinical and
Cochrane Collaboration (3). These pioneering works lead clinical studies, SR and meta-analysis, and CPG.
us to the EBM era. In this era, how to make better use of
preclinical and clinical studies, systematic review (SR) and
meta-analysis, and clinical practice guideline (CPG) attract Methods
extensive attention worldwide. Before using, assessing the
Literature search
methodological quality of study is very important. The qual-
ity includes internal and external validity; methodological We used “methodological quality,” “risk of bias,” and
quality usually refers to internal validity (4, 5). “critical appraisal” to search the Cochrane Handbook (http://
Generally, the medical research can be divided into handbook.cochrane.org/), the Joanna Briggs Institute (JBI)
two major types: primary and secondary. Figure 1 shows Reviewers Manual (http://www.joannabriggs.org/sumari
an overview of the different study types in medical re- .html), and the Centre for Reviews and Dissemination’s
search (6, 7). Different research fields or/and different design guidance for undertaking SRs in health care (http://www
types have different methodological quality assessment tools. .york.ac.uk/inst/crd/index_guidance.htm). The “systematic
The internal validity can be influenced by selection bias, per- review,” meta-analysis, overview, and “clinical practice
formance bias, detection bias, attrition bias, reporting bias, guideline” were also used to search PubMed, we chose the

JEBM 8 (2015) 2–10 


C 2015 Chinese Cochrane Center, West China Hospital of Sichuan University and Wiley Publishing Asia Pty Ltd 3
Methodological quality assessment tools X. Zeng et al.

Table 1 The type and characteristic of methodological quality tools

Type Characteristic

Item Consisting of individual components, relevant to clinical research methodology and the results might be biased, such as allocation
concealment and blinding
Checklist Consisting of many items for assessment study quality and risk of bias, without scoring of each item
Scale Consisting of many items for assessment study quality and risk of bias, every item is scored and combined to give a summary
score

articles that were published in the last five years and extracted Results
the methodological quality tools from them. The “method-
ological quality,” “risk of bias,” “critical appraisal,” validity, Characteristics of tools
tool, item, checklist, and scale were used to search the Google Finally, we included 21 tools in this SR, an overall summary
search engine (http://www.google.com.hk/), we choose the of the main tool characteristics and more detailed information
first 300 links. Reference lists of published articles were is presented in Appendix Table S1 to S8.
examined to identify additional sources not identified in
the database searches. We also searched the website
of the Critical Appraisal Skills Programme (CASP; Tools for randomized controlled trial
http://www.casp-uk.net/), the Scottish Intercollegiate The first randomized controlled trial (RCT) is designed by
Guidelines Network (SIGN; http://www.sign.ac.uk/), and Bradford A. Hill (1897 to 1991) and becomes the “gold
the National Institute for Clinical Excellence (NICE; standard” for experimental study design (9, 10). Archie
http://www.nice.org.uk/). We finished this process on May Cochrane was a staunch supporter of RCT and spent much
20, 2014. of his career promoting their use in research (11) and the
Cochrane Systematic Reviews (CSR) have always focused
on RCT from where it emerged. There are many tools
Eligibility criteria
developed for assessing RCTs. Table S1 presents major items
The eligibility criteria were as follows: (1) the tool for assess- of the Cochrane Collaboration’s tool (5, 12), the Physiother-
ing methodological quality of primary study, SR and meta- apy Evidence Database (PEDro) scale (13–15), the Modified
analysis, or CPG; (2) with the instructions or user’s guide, and Jadad Scale (16–18), the Delphi List (19), CASP checklist for
describes the basic characteristics and objective of the tool; RCT (20–22), and the NICE methodology checklist for RCT.
the tool is also used or commonly used now; (4) the guideline Other tools, such as Chalmers Scale (23), are not com-
for reporting primary study, SR and meta-analysis, or CPG is monly used nowadays, so we did not introduce them. West
excluded; (5) article only provided general narrative guidance et al (24) summarized these tools in 2002.
or without an explicit scale, item, or checklist were excluded.
Tools for nonrandomized interventional
study
Data extraction
In clinical research, especially in surgery, RCT is not always
Two authors independently extracted the following informa- feasible (25). Therefore, nonrandomized designs remain
tion: the type of study addressed by the tool, the type of the considerable. Nonrandomized clinical intervention study
tool (scale, checklist, or item), number of items and major indicates that the investigators control over the allocation of
contents of the tool, and the development unit of the tool. participants to groups, but do not attempt randomization (eg,
If the tool contains internal and external validity (applica- patient or physician preference) (26). According to whether
bility) or other additional criteria at the same time, we only with comparison, nonrandomized clinical intervention study
extracted the section of internal validity. can be divide into comparative and noncomparative. In order
to provide a tool which could be used by readers, manuscript
reviewers or journal editors to assess the quality of such
Data analysis
studies, many scholars developed many methodological
Meta-analysis was inapplicable; the quantitative analysis tools for nonrandomized intervention studies, most of which
method was used to summarize the information and results, can be found in the article by Deeks et al (26).
including objective, version, contents, designer, and website Table S2 presents the Methodological Index for Non-
of tool. randomized Studies (MINORS) tool (27) and Reisch’s

4 JEBM 8 (2015) 2–10 


C 2015 Chinese Cochrane Center, West China Hospital of Sichuan University and Wiley Publishing Asia Pty Ltd
X. Zeng et al. Methodological quality assessment tools

tool (28). The latter tool that was developed by Reisch et al, Cochrane Collaboration, NICE, SIGN, University of York,
and was major for drug therapeutic studies, and is now used and AHRQ (38–42). The Cochrane Collaboration takes
by the Cochrane Inflammatory Bowel Disease Group (In- QUADAS as it is a standard tool for making CSR and set
flammatory Bowel Disease and Functional Bowel Disorders QUADAS in the Review Manager (RevMan) 5.0 software
Review Group) (26). MINORS contain 12 methodological (http://srdta.cochrane.org/handbook-dta-reviews). In 2011,
points, the first eight apply to both noncomparative and a revised “QUADAS-2” is launched due to experience of
comparative studies, while the remaining four relate only to QUADAS Group, reports from users, and feedback from
studies with two or more groups. the Cochrane Collaboration (41, 43). QUADAS-2 tool is set
into RevMan 5.2 in 2012 (44). Another tool is developed by
Tools for analytical study CASP. Table S5 lists the QUADAS, QUADAS-2, and CASP
checklist for assessing the DTA.
Observational study can be divided into analytical study and Other relevant tools are reviewed by Whiting et al in
descriptive study, the analytical study includes three types: 2004 (37).
cohort study, case-control study, and cross-sectional study
(6).
Cohort study includes prospective cohort study, retrospec- Tools for animal study (preclinical study)
tive cohort study, and ambidirectional cohort study (29). Before clinical trials are carried out, the safety and effective-
There are three command tools for assessing the cohort ness of new drugs are usually tested in animal models (45),
study: The CASP checklist, the SIGN methodology tools, therefore, animal studies are considered as preclinical
and the Newcastle-Ottawa Scale (NOS) (30, 31). These three research and have important significance (46, 47). However,
tools (30, 31) also provide criteria for case-control study. unlike clinical studies, the attention on animal studies is far
Nowadays, there are no more accepted tools for cross- from enough. In 2002, the Lancet published a commentary
sectional study (32). The Agency for Healthcare Research that first outlined the scientific rationale for SRs of animal
and Quality (AHRQ) recommends 11 items and the Crom- studies (48) and then many SRs and meta-analyses of animal
bie’s items contains 7 items for assessing the quality of studies were published (47, 49). Like clinical studies, the
cross-sectional/ prevalence study. methodological quality of animal study also needs to be
Table S3 presents these above tools. Other relevant tools assessed (47).
are summarized by Sanderson et al, 2007 (32). Table S6 shows the updated Stroke Therapy Academic
Industry Roundtable (STAIR) tool (50)—Recommendations
Tools for case series for Ensuring Good Scientific Inquiry (51), the revised
A case series is an observational study describing a series of STAIR tool—The Collaborative Approach to Meta-Analysis
individuals, usually all receiving the same intervention with and Review of Animal Data from Experimental Studies
no control group (5). Several tools have been developed for (CAMARADES) (52), and the Systematic Review Centre
assessing the methodological quality of case series study, the for Laboratory Animal Experimentation’s (SYRCLE) risk
latest tool was developed by Moga et al (33) in 2012 using of bias tool (53) for animal studies. These are recommended
a modified Delphi technique, including 18 items. Table S4 tools up to now.
shows this tool.
Other tools for primary study
Tools for diagnostic accuracy study
Qualitative study (qualitative research) design is widely
Diagnostic accuracy study belongs to screening tests; it is used in Nursing. The JBI stipulates that SR must con-
also called “diagnostic test accuracy (DTA).” Screening is duct critical methodological quality using corresponding
defined as tests done among apparently well people to iden- tools; these tools are embed into their premier software—
tify those at an increased risk of a disease or disorder. Those module of JBI System for the Unified Management,
identifications are sometimes then offered a subsequent di- Assessment and Review of Information (SUMARI;
agnostic test or procedure, or in some instances, a treatment http://joannabriggs.org/sumari.html) (54, 55). For quali-
or preventive medication (34, 35). DTA have several unique tative study, JBI gives 10 items and each item is judged
features in terms of design which differ from standard inter- “Yes,” “No,” or “Unclear.” CASP and NICE also provide a
vention and observational evaluations. checklist to assess the methodological quality of qualitative
There are two recommended tools for quality assessment study.
of DTA. The first is called “Quality Assessment of Diag- The JBI, CASP, and NICE also provide tool for assessing
nostic Accuracy Studies” (QUADAS) (36, 37); this tool is the methodological quality of economic evaluations. Besides,
widely used and quickly accepted and recommended by the the CASP provides a tool for appraisal clinical prediction

JEBM 8 (2015) 2–10 


C 2015 Chinese Cochrane Center, West China Hospital of Sichuan University and Wiley Publishing Asia Pty Ltd 5
Methodological quality assessment tools X. Zeng et al.

rule, the NICE provide methodology checklist for prognostic the Appraisal of Guidelines for Research and Evaluation
studies. (AGREE) instrument has the most potential to serve as a
The JBI also gives tool for RCT, quasi-RCT, cohort study, basis for the development of an appraisal tool for clinical
case-control study, cross-sectional study, case report, and pathways (78).
expert opinion, respectively (54, 55). These entire appraisals The AGREE instrument is first released in 2003 (79)
have relevant module of JBI-SUMARI. and updates to AGREE II instrument come in 2009 (80).
AGREE II comprises 23 items to develop, report, and evalu-
ate practice guidelines and is available online at http://www
Tools for SR and meta-analysis .agreetrust.org/resource-centre/. Table S8 presents the major
SR and meta-analysis are the popular methods to keep items of AGREE and AGREE II instrument.
up with current medical literature (56–58). The ultimate
purposes and values of SR and meta-analysis are to use Discussion
them for promoting the health care (58–60). Therefore,
critically appraising the SR and meta-analysis before us- In the EBM era and with the surging number of publica-
ing them, such as conducting overview and evidence-based tions, the major attention is “Going from evidence to rec-
clinical practice, is also necessary and important. SR and ommendations” (81, 82). Critical appraisal of evidence is
meta-analysis assessment include methodological and re- the key point in this process (83, 84). According to charac-
porting quality. The reporting quality is appraised using teristics of different study types, relevant evaluation tools
the relevant guidelines, such as Preferred Reporting Items are developed. Of them, some are recommended and used,
for Systematic reviews and Meta-Analyses (61). The En- some are used without recommendation, and some are elim-
hancing the Quality and Transparency of health Research inated (8, 26), (32), (37–39, 46, 47, 78). These tools give sig-
Network (http://www.equator-network.org/) (62–64) and the nificant impetus to practice of EBM (85, 86).
Research Reporting Guidelines and Initiatives (http://www For primary research, the Collaboration’s recommended
.nlm.nih.gov/services/research_report_guide.html) are good tool for assessing risk of bias of RCT is neither a scale nor a
websites to search reporting guidelines for both primary and checklist; it is a domain-based evaluation in which critical as-
secondary research. sessments are made separately for different domains (5, 12).
In 1988, Sacks et al developed a tool for assessing the qual- The “Cochrane Collaboration’s tool” is very widely ac-
ity of meta-analysis of RCTs and called “Sack’s Quality As- cepted and recommended. Additionally, the Modified Jadad
sessment Check list” (65). And Oxman and Guyatt developed Scale and PEDro Scale are also suitable tools for RCT.
another methodological quality assessment tool—Overview Besides interventional SR of RCTs, the Cochrane Collab-
Quality Assessment Questionnaire (OQAQ) in 1991 (66, 67). oration is now also concerned about SR of DTAs, method-
To overcome the shortcomings of above two tools, a measure- ology, and overview (5). Nonrandomized studies include
ment tool based on them, Assessment of Multiple Systematic quasi-experiments, natural experiments, and observational
reviews (AMSTAR) to assess the methodological quality of studies, which may be prospective or retrospective cohort
SRs is developed in 2007 (68). The AMSTAR tool is an studies or case-control studies (6, 87). Therefore, the
11-item questionnaire that asks reviewers to answer “Yes,” Cochrane Collaboration recommends NOS for assessing
“No,” “Can’t answer,” or “Not applicable” (Table S7). nonrandomized studies and QUADAS and QUADAS-2 for
CASP and NICE also develop a checklist for SR and meta- DTA (5). However, nonrandomized study can be divided into
analysis, Table S7 shows their major items. Besides, the JBI nonrandomized interventional study and observational study.
also sets relevant module of JBI-SUMARI for SR and the Hence, we believe the NOS is applicable to observational
contents are very similar to CASP checklists. study (including case-control and cohort studies) (30, 31).
In this situation, MINORS is defined for nonrandomized
interventional study (27). The aim of Reisch et al (28) was
Clinical practice guideline
to evaluate the design and performance of drug therapeutic
CPG is being well integrated into the thinking of practicing studies, however, that tool lists a total of 57 items grouped
clinicians and professional clinical organizations (69–71); into 12 categories including study design, sample size,
it also incorporates scientific evidence into clinical prac- randomization, and comparison groups (Table S2). Besides,
tice (72). However, not all guidelines are the evidence-based some of the criteria are rather too specific to pharmaceutical
CPGs (73, 74) and the quality of guidelines is of varying studies and would require modification for more general
quality (75–77). Hence, before implementation to clinical, use (26). Therefore, the complicated items and relevant
the quality should be appraised critically. items could be modified limited it use.
In order to improve the quality of guideline and promote to The RCT design is also widely used in animal research
use, more than 20 appraisal tools are developed (78). Of them, and it is absorbing more and more attention (47, 88). Unlike

6 JEBM 8 (2015) 2–10 


C 2015 Chinese Cochrane Center, West China Hospital of Sichuan University and Wiley Publishing Asia Pty Ltd
X. Zeng et al. Methodological quality assessment tools

clinical RCT, the methodological assessment tools for clini- well-accepted tools remains a significant and important
cal RCT cannot be used in animal study. In 1999, the initial work.
“STAIR” group recommendations their criteria for assessing In addition, it was important that all assessment tools
the quality of stroke animal studies and hence this tool is also with one common deficit is subjective. Therefore, the user
called “STAIR.” In 2009, the STAIR Group updates their cri- must receive formal training, with relevant epidemiological
teria and developed “Recommendations for Ensuring Good knowledge, and have rigorous academic attitude. At least
Scientific Inquiry.” Macleod et al proposed a 10-points tool— two reviewers independently evaluating and cross-checking
CAMARADES based on STAIR to assess methodological is a good method to avoid performance bias (93). Second,
quality of animal study in 2004 (52). The SYRCLE’s RoB the reporting quality is not equal to methodological quality,
tool is developed based on the Cochrane risk of bias tool in hence using reporting guidelines/statements to assess
2014 (53). The first two appraisal tools are both major for methodological quality is not appropriate.
stroke animal studies and they are accepted and used in stroke In conclusion, making accurate judgment of study type
animal experiments (89–92). Although the CAMARADES is the first priority. Then, choosing the corresponding tool
tool gives relevant modified items for other animals besides is also important. In one word that mastering the relevant
stroke, the public acceptance is not very high. knowledge comprehensively and with more practice is the
JBI releases the most types of tools, including RCT, basic requirement for assessing the methodological quality
quasi-RCT, cohort study, case-control study, nonexperimen- correctly.
tal study (descriptive study), case report, expert opinion,
qualitative study, and SR (54, 55). However, their use is not Funding
too much and in the nursing field. The reason might be as
it requires that the user must have permission and major This research was supported (in part) by the National Scien-
concerns on nursing field. We consider that the JBI tools are tific Fund and project (81302508, 81403276), China Medi-
well and can be used in relevant studies. cal Board Collaboration Program in Health Policy Evidence
CASP has many checklists and every one consists of three (12-095), Foundation of Education and Science Planning
parts: “Are the results of the trial valid” (Section A), “What Project of Hubei Province (2012A050), the special funds
are the results” (Section B), and “Will the results help lo- for “985 project” construction project of Wuhan University,
cally” (Section C). For evaluating the methodological qual- and the Intramural Research Program of the Hubei University
ity, we use section A the most. The sections B and C are of Medicine (2011CZX01), without commercial or not-for-
for evidence-based practice. CASP checklist for “Clinical profit sectors. The funders had no role in study design, data
Prediction Rule” is the only tool in this field. collection and analysis, decision to publish, or preparation of
NICE provides seven methodology checklists: SR and the manuscript. No additional external funding was received
meta-analysis, RCT, cohort study, case-control study, eco- for this study.
nomic evaluation, qualitative study, and prognostic study;
and recommends QUADAS-2 tool for DTA. References
Overview is designed to compile evidence from multiple
SRs of a topic into one accessible and usable document (5). 1. Stavrou A, Challoumas D, Dimitrakakis G. Archibald Cochrane
For SR and meta-analysis, besides above mentioned JBI tool, (1909–1988): the father of evidence-based medicine. Interactive
Cardiovascular and Thoracic Surgery 2013; 18(1): 121–24.
CASP and NICE checklists, AMSTAR tool (68) is the ac-
2. Group E-BMW. Evidence-based medicine. A new approach to
cepted and recommended tool. The QQAQ is also widely
teaching the practice of medicine. JAMA 1992; 268(17):
used nowadays. For CPG, the AGREE II instrument (80) is
2420–25.
a well-accepted and recommended tool.
3. Levin A. The Cochrane Collaboration. Annals of Internal
However, great efforts remain necessary in the appraisal
Medicine 2001; 135(4): 309–12.
tools. First, many research types lack assessment tools, such 4. Campbell DT. Factors relevant to the validity of experiments in
as before-after study (time series) and nested case-control social settings. Psychological Bulletin 1957; 54(4): 297–312.
study (29). Second, special field lack corresponding tool, 5. Higgins J, Green S. Cochrane Handbook for Systematic Reviews
such as genetic study (known as single-nucleotide polymor- of Interventions Version 5.1.0 [updated March 2011]: The
phism) and in vivo experiment (cell study). Especially of Cochrane Collaboration, 2011.
single-nucleotide polymorphism research, the meta-analysis 6. Grimes DA, Schulz KF. An overview of clinical research: the lay
of this aspect is too many but few of them perform of the land. Lancet 2002; 359(9300): 57–61.
methodological quality assessment. Third, the existing tools 7. Rohrig B, du Prel JB, Wachtlin D, Blettner M. Types of study in
are not well accepted, such as the tools for cross-sectional medical research: part 3 of a series on evaluation of scientific
study and animal study. In further study, how to develop publications. Dtsch Arztebl Int 2009; 106(15): 262–68.

JEBM 8 (2015) 2–10 


C 2015 Chinese Cochrane Center, West China Hospital of Sichuan University and Wiley Publishing Asia Pty Ltd 7
Methodological quality assessment tools X. Zeng et al.

8. Juni P, Altman DG, Egger M. Systematic reviews in health care: 24. West S, King V, Carey TS, Lohr KN, McKoy N, Sutton SF, et al.
assessing the quality of controlled clinical trials. BMJ 2001; Systems to rate the strength of scientific evidence. Evidence
323(7303): 42–46. Report-Technology Assessment (Summary) 2002; 47: 1–11.
9. The British Tuberculosis Association Research Committee. 25. McCulloch P, Taylor I, Sasako M, Lovett B, Griffin D.
Treatment of pulmonary tuberculosis with streptomycin and Randomised trials in surgery: problems and possible solutions.
para-aminosalicylic acid: a Medical Research Council BMJ 2002; 324(7351): 1448–51.
investigation. British Medical Journal 1950; 2(4688): 1073–85. 26. Deeks JJ, Dinnes J, D’Amico R, Sowden AJ, Sakarovitch C,
10. Armitage P. Fisher, Bradford Hill, and randomization. Song F, et al. Evaluating non-randomised intervention studies.
International Journal of Epidemiology 2003; 32(6): 925–28; Health Technology Assessment 2003; 7(27): 1–173.
discussion 45–48. 27. Slim K, Nini E, Forestier D, Kwiatkowski F, Panis Y, Chipponi
11. Shah HM, Chung KC. Archie Cochrane and his vision for J. Methodological index for non-randomized studies (minors):
evidence-based medicine. Plastic and Reconstructive Surgery development and validation of a new instrument. ANZ Journal of
2009; 124(3): 982–88. Surgery 2003; 73(9): 712–16.
12. Higgins JP, Altman DG, Gotzsche PC, Juni P, Moher D, Oxman 28. Reisch JS, Tyson JE, Mize SG. Aid to the evaluation of
AD, et al. The Cochrane Collaboration’s tool for assessing risk therapeutic studies. Pediatrics 1989; 84(5): 815–27.
of bias in randomised trials. BMJ 2011; 343: d5928. 29. Grimes DA, Schulz KF. Cohort studies: marching towards
13. Campos TF, Beckenkamp PR, Moseley AM. Usage evaluation outcomes. Lancet 2002; 359(9303): 341–45.
of a resource to support evidence-based physiotherapy: the 30. Wells G, Shea B, O’Connell D, Peterson J, Welch V, Losos M,
physiotherapy evidence database (PEDro). Physiotherapy 2013; et al. The Newcastle-Ottawa Scale (NOS) for Assessing the
99(3): 252–57. Quality of Nonrandomised Studies in Meta-Analyses.
14. Maher CG, Sherrington C, Herbert RD, Moseley AM, Elkins M. http://www.ohri.ca/programs/clinical_epidemiology/
Reliability of the PEDro scale for rating quality of randomized oxford.asp Accessed in June, 2014.
controlled trials. Physical Therapy 2003; 83(8): 713–21. 31. Stang A. Critical evaluation of the Newcastle-Ottawa scale for
15. Shiwa SR, Costa LO, Costa Lda C, Moseley A, Hespanhol, LC the assessment of the quality of nonrandomized studies in
Jr, Venancio R, et al. Reproducibility of the Portuguese version meta-analyses. European Journal of Epidemiology 2010; 25(9):
of the PEDro scale. Cad Saude Publica 2011; 27(10): 2063–68. 603–05.
16. Jadad AR, Moore RA, Carroll D, Jenkinson C, Reynolds DJ, 32. Sanderson S, Tatt ID, Higgins JP. Tools for assessing quality and
Gavaghan DJ, et al. Assessing the quality of reports of susceptibility to bias in observational studies in epidemiology: a
randomized clinical trials: is blinding necessary? Controlled systematic review and annotated bibliography. International
Clinical Trials 1996; 17(1): 1–12. Journal of Epidemiology 2007; 36(3): 666–76.
17. Schulz KF, Chalmers I, Hayes RJ, Altman DG. Empirical 33. Moga C, Guo B, Schopflocher D, Harstall C. Development of a
evidence of bias. Dimensions of methodological quality Quality Appraisal Tool for Case Series Studies using a Modified
associated with estimates of treatment effects in controlled trials. Delphi Technique. 2012. http://www.ihe.ca/documents/Case%20
JAMA 1995; 273(5): 408–12. series%20studies%20using%20a%20modified%20Delphi%20
18. Hartling L, Ospina M, Liang Y, Dryden DM, Hooton N, Krebs technique.pdf Accessed in June, 2014.
Seida J, et al. Risk of bias versus quality assessment of 34. Grimes DA, Schulz KF. Uses and abuses of screening tests.
randomised controlled trials: cross sectional study. BMJ 2009; Lancet 2002; 359(9309): 881–84.
339: b4012. 35. Grimes DA, Schulz KF. Refining clinical diagnosis with
19. Verhagen AP, de Vet HC, de Bie RA, Kessels AG, Boers M, likelihood ratios. Lancet 2005; 365(9469): 1500–05.
Bouter LM, et al. The Delphi list: a criteria list for quality 36. Whiting P, Rutjes AW, Reitsma JB, Bossuyt PM, Kleijnen J.
assessment of randomized clinical trials for conducting The development of QUADAS: a tool for the quality assessment
systematic reviews developed by Delphi consensus. Journal of of studies of diagnostic accuracy included in systematic reviews.
Clinical Epidemiology 1998; 51(12): 1235–41. BMC Medical Research Methodology 2003; 3: 25.
20. Ibbotson T, Grimshaw J, Grant A. Evaluation of a programme of 37. Whiting P, Rutjes AW, Dinnes J, Reitsma J, Bossuyt PM,
workshops for promoting the teaching of critical appraisal skills. Kleijnen J. Development and validation of methods for assessing
Medical Education 1998; 32(5): 486–91. the quality of diagnostic accuracy studies. Health Technology
21. Singh J. Critical appraisal skills programme. Journal of Assessment 2004; 8(25): 1–234.
Pharmacology and Pharmacotherapeutics 2013; 4(1): 76. 38. Willis BH, Quigley M. Uptake of newer methodological
22. Taylor R, Reeves B, Ewings P, Binns S, Keast J, Mears R. A developments and the deployment of meta-analysis in diagnostic
systematic review of the effectiveness of critical appraisal skills test research: a systematic review. BMC Medical Research
training for clinicians. Medical Education 2000; 34(2): 120–25. Methodology 2011; 11: 27.
23. Chalmers TC, Smith H Jr., Blackburn B, Silverman B, 39. Whiting PF, Rutjes AW, Westwood ME, Mallett S, Group Q-S.
Schroeder B, Reitman D, et al. A method for assessing the A systematic review classifies sources of bias and variation in
quality of a randomized control trial. Controlled Clinical Trials diagnostic test accuracy studies. Journal of Clinical
1981; 2(1): 31–49. Epidemiology 2013; 66(10): 1093–1104.

8 JEBM 8 (2015) 2–10 


C 2015 Chinese Cochrane Center, West China Hospital of Sichuan University and Wiley Publishing Asia Pty Ltd
X. Zeng et al. Methodological quality assessment tools

40. Whiting PF, Weswood ME, Rutjes AW, Reitsma JB, Bossuyt 56. Lau J, Ioannidis JP, Schmid CH. Summing up evidence: one
PN, Kleijnen J. Evaluation of QUADAS, a tool for the quality answer is not always enough. Lancet 1998; 351(9096): 123–27.
assessment of diagnostic accuracy studies. BMC Medical 57. Clarke M, Chalmers I. Meta-analyses, multivariate analyses, and
Research Methodology 2006; 6: 9. coping with the play of chance. Lancet 1998; 351(9108):
41. Whiting PF, Rutjes AW, Westwood ME, Mallett S, Deeks JJ, 1062–63.
Reitsma JB, et al. QUADAS-2: a revised tool for the quality 58. Oxman AD, Schunemann HJ, Fretheim A. Improving the use of
assessment of diagnostic accuracy studies. Annals of Internal research evidence in guideline development: 8. Synthesis and
Medicine 2011; 155(8): 529–36. presentation of evidence. Health Research Policy and Systems
42. Deeks J, Bossuyt P, Gatsonis C. Cochrane Handbook for 2006; 4: 20.
Systematic Reviews of Diagnostic Test Accuracy Version 1.0.0. 59. Swennen MH, van der Heijden GJ, Boeije HR, van Rheenen N,
2009. http://srdta.cochrane.org/ Accessed in June, 2014. Verheul FJ, van der Graaf Y, et al. Doctors’ perceptions and use
43. Schueler S, Schuetz GM, Dewey M. The revised QUADAS-2 of evidence-based medicine: a systematic review and thematic
tool. Annals of Internal Medicine 2012; 156(4): 323; author synthesis of qualitative studies. Academic Medicine 2013; 88(9):
reply 4. 1384–96.
44. Review Manager (RevMan). Version 5.2 [program]. 60. Gallagher EJ. Systematic reviews: a logical methodological
Copenhagen: The Nordic Cochrane Centre, The Cochrane extension of evidence-based medicine. Academic Emergency
Collaboration, 2012. Medicine 1999; 6(12): 1255–60.
45. Sibbald WJ. An alternative pathway for preclinical research in 61. Liberati A, Altman DG, Tetzlaff J, Mulrow C, Gotzsche PC,
fluid management. Critical Care 2000; 4 (Suppl2): Ioannidis JP, et al. The PRISMA statement for reporting
S8–S15. systematic reviews and meta-analyses of studies that evaluate
46. Perel P, Roberts I, Sena E, Wheble P, Briscoe C, Sandercock P, healthcare interventions: explanation and elaboration. BMJ
et al. Comparison of treatment effects between animal 2009; 339.
experiments and clinical trials: systematic review. BMJ 2007; 62. Harms M. The EQUATOR network and the PRISMA statement
334(7586): 197. for the reporting of systematic reviews and meta-analyses.
47. Hooijmans CR, Ritskes-Hoitinga M. Progress in using Physiotherapy 2009; 95(4): 237–40.
systematic reviews of animal studies to improve translational 63. Pandis N, Fedorowicz Z. The international EQUATOR network:
research. PLoS Medicine 2013; 10(7): e1001482. enhancing the quality and transparency of health care research.
48. Sandercock P, Roberts I. Systematic reviews of animal Journal of Applied Oral Science 2011; 19(5): doi:
experiments. Lancet 2002; 360(9333): 586. pii:S1678-77572011000500001.
49. van Luijk J, Leenaars M, Hooijmans C, Wever K, de Vries R, 64. Kac G, Hirst A. Enhanced quality and transparency of health
Ritskes-Hoitinga M. Towards evidence-based translational research reporting can lead to improvements in public health
research: the pros and cons of conducting systematic reviews of policy decision-making: help from the EQUATOR Network.
animal studies. ALTEX 2013; 30(2): 256–57. Cad Saude Publica 2011; 27(10): 1872.
50. Stroke Therapy Academic Industry R. Recommendations for 65. Sacks HS, Berrier J, Reitman D, Ancona-Berk VA, Chalmers
standards regarding preclinical neuroprotective and restorative TC. Meta-analyses of randomized controlled trials. New
drug development. Stroke 1999; 30(12): 2752–58. England Journal of Medicine 1987; 316(8): 450–55.
51. Fisher M, Feuerstein G, Howells DW, Hurn PD, Kent TA, Savitz 66. Oxman AD. Checklists for review articles. BMJ 1994;
SI, et al. Update of the stroke therapy academic industry 309(6955): 648–51.
roundtable preclinical recommendations. Stroke 2009; 40(6): 67. Oxman AD, Guyatt GH. Validation of an index of the quality of
2244–50. review articles. Journal of Clinical Epidemiology 1991; 44(11):
52. Macleod MR, O’Collins T, Howells DW, Donnan GA. Pooling 1271–78.
of animal experimental data reveals influence of study design 68. Shea BJ, Grimshaw JM, Wells GA, Boers M, Andersson N,
and publication bias. Stroke 2004; 35(5): 1203–08. Hamel C, et al. Development of AMSTAR: a measurement tool
53. Hooijmans CR, Rovers MM, de Vries RB, Leenaars M, to assess the methodological quality of systematic reviews. BMC
Ritskes-Hoitinga M, Langendam MW. SYRCLE’s risk of bias Medical Research Methodology 2007; 7: 10.
tool for animal studies. BMC Medical Research Methodology 69. Davis DA, Taylor-Vaisey A. Translating guidelines into practice.
2014; 14: 43. A systematic review of theoretic concepts, practical experience
54. Vardell E, Malloy M. Joanna Briggs Institute: an evidence-based and research evidence in the adoption of clinical practice
practice database. Medical Reference Services Quarterly 2013; guidelines. Canadian Medical Association Journal 1997;
32(4): 434–42. 157(4): 408–16.
55. Hannes K, Lockwood C. Pragmatism as the philosophical 70. Neely JG, Graboyes E, Paniello RC, Sequeira SM, Grindler DJ.
foundation for the Joanna Briggs meta-aggregative approach to Practical guide to understanding the need for clinical practice
qualitative evidence synthesis. Journal of Advanced Nursing guidelines. Otolaryngology–Head and Neck Surgery 2013;
2011; 67(7): 1632–42. 149(1): 1–7.

JEBM 8 (2015) 2–10 


C 2015 Chinese Cochrane Center, West China Hospital of Sichuan University and Wiley Publishing Asia Pty Ltd 9
Methodological quality assessment tools X. Zeng et al.

71. Browman GP, Levine MN, Mohide EA, Hayward RS, Pritchard 82. Andrews J, Guyatt G, Oxman AD, Alderson P, Dahm P,
KI, Gafni A, et al. The practice guidelines development cycle: a Falck-Ytter Y, et al. GRADE guidelines: 14. Going from
conceptual tool for practice guidelines development and evidence to recommendations: the significance and presentation
implementation. Journal of Clinical Oncology 1995; 13(2): of recommendations. Journal of Clinical Epidemiology 2013;
502–12. 66(7): 719–25.
72. Tracy SL. From bench-top to chair-side: how scientific evidence 83. Tunguy-Desmarais GP. Evidence-based medicine should be
is incorporated into clinical practice. Dental Materials 2013; based on science. South African Medical Journal 2013; 103(10):
30(1): 1–15. 700.
73. Chapa D, Hartung MK, Mayberry LJ, Pintz C. Using 84. Muckart DJ. Evidence-based medicine - are we boiling the frog?
preappraised evidence sources to guide practice decisions. South African Medical Journal 2013; 103(7): 447–48.
Journal of the American Association of Nurse Practitioners 85. Swanson JA, Schmitz D, Chung KC. How to practice
2013; 25(5): 234–43. evidence-based medicine. Plastic and Reconstructive Surgery
74. Eibling D, Fried M, Blitzer A, Postma G. Commentary on the 2010; 126(1): 286–94.
role of expert opinion in developing evidence-based guidelines. 86. Manchikanti L. Evidence-based medicine, systematic reviews,
Laryngoscope 2013; 124(2): 355–57. and guidelines in interventional pain management, part I:
75. Chen YL, Yao L, Xiao XJ, Wang Q, Wang ZH, Liang FX, et al. introduction and general considerations. Pain Physician 2008;
Quality assessment of clinical guidelines in China: 1993 - 2010. 11(2): 161–86.
Chinese Medical Journal (English) 2012; 125(20): 3660–64. 87. Britton A, McKee M, Black N, McPherson K, Sanderson C,
76. Hu J, Chen R, Wu S, Tang J, Leng G, Kunnamo I, et al. The Bain C. Choosing between randomised and non-randomised
quality of clinical practice guidelines in China: a systematic studies: a systematic review. Health Technology Assessment
assessment. Journal of Evaluation in Clinical Practice 2013; 1998; 2(13): 1–124.
19(5): 961–67. 88. Vesterinen HM, Sena ES, Egan KJ, Hirst TC, Churolov L, Currie
77. Henig O, Yahav D, Leibovici L, Paul M. Guidelines for the GL, et al. Meta-analysis of data from animal studies: a practical
treatment of pneumonia and urinary tract infections: evaluation guide. Journal of Neuroscience Methods 2013; 221C: 92–102.
of methodological quality using the Appraisal of Guidelines, 89. O’Connor AM, Sargeant JM. Meta-analyses including data from
Research and Evaluation II instrument. Clinical Microbiology observational studies. Preventive Veterinary Medicine 2013;
and Infection 2013; 19(12): 1106–14. 113(3): 313–22.
78. Vlayen J, Aertgeerts B, Hannes K, Sermeus W, Ramaekers D. A 90. Schmidt A, Wellmann J, Schilling M, Strecker JK, Sommer C,
systematic review of appraisal tools for clinical practice Schabitz WR, et al. Meta-analysis of the efficacy of different
guidelines: multiple similarities and one common deficit. training strategies in animal models of ischemic stroke. Stroke
International Journal for Quality in Health Care 2005; 17(3): 2013; 45(1): 239–47.
235–42. 91. Wei RL, Teng HJ, Yin B, Xu Y, Du Y, He FP, et al. A systematic
79. Collaboration A. Development and validation of an international review and meta-analysis of Buyang Huanwu decoction in
appraisal instrument for assessing the quality of clinical practice animal model of focal cerebral ischemia. Evidence-Based
guidelines: the AGREE project. Quality & Safety in Health Care Complementary and Alternative Medicine 2013, 2013: 138484.
2003; 12(1): 18–23. 92. Wu S, Sena E, Egan K, Macleod M, Mead G. Edaravone
80. Brouwers MC, Kho ME, Browman GP, Burgers JS, Cluzeau F, improves functional and structural outcomes in animal models of
Feder G, et al. AGREE II: advancing guideline development, focal cerebral ischemia: a systematic review. International
reporting and evaluation in health care. Canadian Medical Journal of Stroke 2013; 9(1): 101–06.
Association Journal 2010; 182(18): E839–E842. 93. Gold C, Erkkila J, Crawford MJ. Shifting effects in randomised
81. Guyatt GH, Oxman AD, Kunz R, Falck-Ytter Y, Vist GE, controlled trials of complex interventions: a new kind of
Liberati A, et al. Going from evidence to recommendations. performance bias? Acta Psychiatrica Scandinavica 2012;
BMJ 2008; 336(7652): 1049–51. 126(5): 307–14.

10 JEBM 8 (2015) 2–10 


C 2015 Chinese Cochrane Center, West China Hospital of Sichuan University and Wiley Publishing Asia Pty Ltd

You might also like