Professional Documents
Culture Documents
Paul Cronin, MD, MS, James V. Rawson, MD, Marta E. Heilbrun, MD, MS, Janie M. Lee, MD, MSc,
Aine M. Kelly, MD, MS, MA, Pina C. Sanelli, MD, MPH, Brian W. Bresnahan, PhD, Angelisa M. Paladin, MD
Incomplete reporting hampers the evaluation of results and bias in clinical research studies. Guidelines for reporting study design and
methods have been developed to encourage authors and journals to include the required elements. Recent efforts have been made to
standardize the reporting of clinical health research including clinical guidelines. In this article, the reporting of diagnostic test accuracy
studies, screening studies, therapeutic studies, systematic reviews and meta-analyses, cost-effectiveness assessments (CEA), recommendations and/or guidelines, and medical education studies is discussed. The available guidelines, many of which can be found at
the Enhancing the QUAlity and Transparency Of health Research network, on how to report these different types of health research are
also discussed. We also hope that this article can be used in academic programs to educate the faculty and trainees of the available resources to improve our health research.
Key Words: Cost-effectiveness assessments; diagnostic test accuracy; enhancing the quality and transparency of health research;
guidelines; medical education; recommendations; reporting; screening; systematic reviews and meta-analyses; therapy.
AUR, 2014
1088
Item
1
Title/Abstract/Key words
Identify the article as a study of diagnostic
accuracy
Introduction
State the research questions or study aims, such as
estimating diagnostic accuracy or comparing
accuracy between tests or across participant groups.
Methods Participants
Describe the study population: The inclusion and
exclusion criteria, setting and locations where
data were collected.
Describe participant recruitment and sampling: How
eligible patients are identified.
10
Describe the scientific background, previous work on the subject, the remaining uncertainty, and, hence, the
rationale for their study.
Clearly specified research questions help the readers to judge the appropriateness of the study design and data
analysis.
Diagnostic accuracy studies describe the behavior of a test under particular circumstances and should report its
inclusion and exclusion criteria for selecting the study population. The spectrum of the target disease can vary
and affect test performance.
Was recruitment based on presenting symptoms, results from previous tests, or the fact that the participants had
received the index tests or the reference standard?
Describe how eligible subjects were identified and whether the study enrolls consecutive or random sampling of
patients. Study designs are likely to influence the spectrum of disease represented.
Was the study population a consecutive series of participants defined by the selection criteria in item 3 and 4? If
not, specify how participants were further selected.
Prospective data collection has many advantages: better data control, additional checks for data integrity and
consistency, and a level of clinical detail appropriate to the problem. As a result, there will be fewer missing or
uninterpretable data items.
Retrospective data collection starts after patients have undergone the index test and the reference standard and
often relies on chart review. Studies with retrospective data collection may reflect routine clinical practice
better than a prospective study, but also may fail to identify all eligible patients or to provide data of high quality.
The reference standard is used to distinguish patients with and without disease. When it is not possible to subject
all patients to the reference standard for practical or ethical reasons, composite reference standard is an
alternative. The components may reflect different definitions or strategies for disease diagnosis.
Describe the methods involved in the execution of index test and reference standard in sufficient detail to allow
other researchers to replicate the study. Differences in the execution of the index test and reference standard
are a potential source of variation in diagnostic accuracy.
The description should cover the full test protocol including the specification of materials and instruments
together with their instructions for use.
Test results can be truly dichotomous (eg, present or absent), have multiple categories or be continuous. Clearly
describe how and when category boundaries are used.
1089
Variability in the manipulation, processing, or reading of the index test or reference standard will affect diagnostic
accuracy.
Professional background, expertise, and prior training to improve interpretation and to reduce interobserver
variation all affect the quality of reading.
(Continued on next page)
Item
11
12
13
14
15
16
17
19
20
21
Results Participants
Report when study was performed, including
beginning and end dates of recruitment.
Report Clinical and demographic characteristics of
the study population.
Report the number of participants satisfying the
criteria for inclusion who did or did not undergo
the index tests and/or the reference standard.
Results Test results
Report time interval between the index tests and
the reference standard, and any treatment
administered in between.
Report distribution of severity of disease (define
criteria).
Knowledge of the results of the reference standard can influence the reading of the index test, and vice versa,
leading to inflated measures of diagnostic accuracy.
Sensitivity, specificity, PPV, NPV, ROC, likelihood ratio and odds ratio.
Reproducibility of the index test and reference standard varies. Poor reproducibility adversely affects diagnostic
accuracy. If possible, authors should evaluate the reproducibility of the test methods used in their study and
report their procedure to do so.
Technology behind many tests advances continuously, leading to improvements in diagnostic accuracy. There
may be a considerable gap between the dates of the study and the publication date of the study report.
Description of the demographic and clinical characteristics are usually presented in a table, such as age, sex,
spectrum of presenting symptoms, comorbidity, current treatments, recruitment centers.
Describe why participants failed to receive either test.
Flow diagram is strongly recommended.
When delay occurs between doing the index test and the reference standard the condition of the patient may
change, leading to worsening or improvement of the disease. Similar concerns apply if treatment is started after
doing the index test but before doing the reference standard.
Demographic and clinical features of the study population can affect measures of diagnostic accuracy. Many
diseases are not pure dichotomous states but cover a continuum, ranging from minute pathological changes to
advanced clinical disease. Test sensitivity is often higher in studies with a higher proportion of patients with
more advanced stages of the target condition.
Cross tabulations of test results in categories and graphs of distributions of continuous results are essential to
allow scientific colleagues to (re)calculate measures of diagnostic accuracy or to perform alternative analyses,
including meta-analysis.
Not all tests are safe. Measuring and reporting of adverse events in studies of diagnostic accuracy can provide
additional information about the clinical usefulness of a particular test.
Report a value of how well the test results correspond with the reference standard. The values presented in the
report should be taken as estimates with some variation. Many journals require or strongly encourage the use of
confidence intervals as measures of precision. A 95% confidence interval is conventional.
18
CRONIN ET AL
1090
TABLE 2. Checklist for Reporting the Reference Case CostEffectiveness Analysis (61)
NPV, negative predictive value; PPV, positive predictive value; ROC, receiver operating characteristic.
Provide a general interpretation of the results in the context of current evidence and its applicability in practice.
Clearly define the methodological shortcomings of the study, how it potentially affected the results and
approaches to limit its significance.
Discuss differences between the context of the study and other settings and patient groups in which the test is
likely to be used.
Provide future direction for this work in advancing clinical practice or research in this field.
Discussion
Discuss the clinical applicability of the study findings.
25
24
22
Uninterpretable, indeterminate, and intermediate test results pose a problem in the assessment of a diagnostic
test. By itself, the frequency of these test results is an important indicator of the overall usefulness of the test.
Furthermore, ignoring such test results can produce biased estimates of diagnostic accuracy if these results
occur more frequently in patients with disease than in those without, or vice versa.
Since variability is the rule rather than the exception, researchers should explore possible sources of
heterogeneity in results, within the limits of the available sample size. The best practice is to plan subgroup
analyses before the start of the study.
Report all measures of test reproducibility performed during the study. For quantitative analytical methods,
report the coefficient of variation (CV).
Framework
1
Background of the problem
2
General framing and design of the analysis
3
Target population for intervention
4
Other program descriptors (eg, care setting, model of
delivery, timing of intervention)
5
Description of comparator programs
6
Boundaries of the analysis
7
Time horizon
8
Statement of the perspective of the analysis
Data and Methods
9
Description of event pathway
10
Identification of outcomes of interest in analysis
11
Description of model used
12
Modeling assumptions
13
Diagram of event pathway (model)
14
Software used
15
Complete description of estimates of effectiveness,
resource use, unit costs, health states, and
quality-of-life weights and their sources
16
Methods for obtaining estimates of effectiveness, costs,
and preferences
17
Critique of data quality
18
Statement of year of costs
19
Statement of method used to adjust costs for inflation
20
Statement of type of currency
21
Source and methods for obtaining expert judgment
22
Statement of discount rates
Results
23
Results of model validation
24
Reference case results (discounted at 3% and
undiscounted): total costs and effectiveness,
incremental costs and effectiveness, and incremental
cost-effectiveness ratios
25
Results of sensitivity analyses
26
Other estimates of uncertainty, if available
27
Graphical representation of cost-effectiveness results
28
Aggregate cost and effectiveness Information
29
Disaggregated results, as relevant
30
Secondary analyses using 5% discount rate
31
Other secondary analyses, as relevant
Discussion
32
Summary of reference case results
33
Summary of sensitivity of results to assumptions and
uncertainties in the analysis
34
Discussion of analysis assumptions having important
ethical implications
35
Limitations of the study
36
Relevance of study results for specific policy questions
or decisions
37
Results of related cost-effectiveness analyses
38
Distributive implications of an intervention
CRONIN ET AL
TABLE 3. International Society for Pharmacoeconomics and Outcomes Research Randomized Control Trial Cost-Effectiveness
Analysis (ISPOR RCT-CEA) Task Force Report of Core Recommendations for Conducting Economic Analyses Alongside Clinical
Trials (62)
Trial design
1
Trial design should reflect effectiveness rather than efficacy when possible.
2
Full follow-up of all patients is encouraged.
3
Describe power and ability to test hypotheses, given the trial sample size.
4
Clinical end points used in economic evaluations should be disaggregated.
5
Direct measures of outcome are preferred to use of intermediate end points.
Data elements
6
Obtain information to derive health state utilities directly from the study population.
7
Collect all resources that may substantially influence overall costs; these include those related and unrelated to the intervention.
Database design and management
8
Collection and management of the economic data should be fully integrated into the clinical data.
9
Consent forms should include wording permitting the collection of economic data, particularly when it will be gathered from thirdparty databases and may include pre- and/or post-trial records.
Analysis
10
The analysis of economic measures should be guided by a data analysis plan and hypotheses that are drafted prior to the onset of
the study.
11
All cost-effectiveness analyses should include the following: an intention-to-treat analysis; common time horizon(s) for
accumulating costs and outcomes; a within-trial assessment of costs and outcomes; an assessment of uncertainty; a common
discount rate applied to future costs and outcomes; an accounting for missing and/or censored data.
12
Incremental costs and outcomes should be measured as differences in arithmetic means, with statistical testing accounting for
issues specific to these data (eg, skewness, mass at zero, censoring, construction of QALYs).
13
Imputation is desirable if there is a substantial amount of missing data. Censoring, if present, should also be addressed.
14
One or more summary measures should be used to characterize the relative value of the intervention.
15
Examples include ratio measures, difference measures, and probability measures (eg, cost-effectiveness acceptability curves).
16
Uncertainty should be characterized. Account for uncertainty that stems from sampling, fixed parameters such as unit costs and
the discount rate, and methods to address missing data.
17
Threats to external validityincluding protocol-driven resource use, unrepresentative recruiting centers, restrictive inclusion and
exclusion criteria, and artificially enhanced complianceare best addressed at the design phase.
18
Multinational trials require special consideration to address intercountry differences in population characteristics and treatment
patterns.
19
When models are used to estimate costs and outcomes beyond the time horizon of the trial, good modeling practices should be
followed.
Models should reflect the expected duration of the intervention on costs and outcomes.
20
Subgroup analyses based on prespecified clinical and economic interactions, when found to be significant ex post, are
appropriate. Ad hoc subgroup analysis is discouraged.
Reporting the results
21
Minimum reporting standards for cost-effectiveness analyses should be adhered to for those conducted alongside clinical trials.
22
The cost-effectiveness report should include a general description of the clinical trial and key clinical findings.
23
Reporting should distinguish economic data collected as part of the trial vs. data not collected as part of the trial.
24
The amount of missing data should be reported. If imputation methods are used, the method should be described.
25
Methods used to construct and compare costs and outcomes, and to project costs and outcomes beyond the trial period should
be described.
26
The results section should include summaries of resource use, costs, and outcome measures, including point estimates and
measures of uncertainty. Results should be reported for the time horizon of the trial, and for projections beyond the trial (if
conducted).
27
Graphical displays are recommended for results not easily reported in tabular form (eg, cost-effectiveness acceptability curves,
joint density of incremental costs and outcomes).
QALYs, quality-adjusted life years.
Establishing transparency
Management of conflict of interest
Guideline development group composition
Clinical practice guideline-systematic review
intersection
Establishing evidence foundations for and rating
strength of recommendations
Articulations of recommendations
External review
Updating
Based on clinical practice guidelines we can trust, Institute of Medicine, National Academic Press, 2011 (64).
improvement studies; consolidated health economic evaluation reporting standards (CHEERS) for health economics
studies (3039); and statement on reporting of evaluation
studies in health informatics (STARE-HI) for studies of
health informatics.
HOW TO REPORT STUDIES
Screening Studies and Diagnostic Test Accuracy
Studies
Research
Study Design
Reporting Guidelines
Provided For
Reporting Guideline
Acronym
Full Text if
Available
Studies of diagnostic
accuracy
STARD
http://www.stardstatement.org/
CONSORT
http://www.consortstatement.org/
Trials assessing
nonpharmacologic
treatments
CONSORT
nonpharmacological
treatment
interventions
http://www.consortstatement.org/
extensions/interventions/
non-pharmacologictreatment-interventions/
CRONIN ET AL
1094
TABLE 5. Reporting Guidelines by Research Study Design, Acronym, Web site URL, and Bibliographic Reference
CONSORT pragmatic
trials
http://www.consortstatement.org/
extensions/designs/
pragmatic-trials/
Reporting of harms in
randomized trials
CONSORT Harms
http://www.consortstatement.org/
extensions/data/harms/
Patient-reported outcomes
in randomized trials
CONSORT-PRO
http://www.consortstatement.org/
extensions/data/pro/
Reporting of noninferiority
and equivalence
randomized trials
CONSORT noninferiority
http://www.consortstatement.org/
extensions/designs/noninferiority-andequivalence-trials/
SPIRIT
http://www.spiritstatement.org/
PRISMA
http://www.prismastatement.org/
1095
CONSORT Cluster
Systematic reviews/
meta-analyses/HTA
http://www.consortstatement.org/
extensions/designs/
cluster-trials/
http://www.consortstatement.org/
extensions/data/
abstracts/
Research
Study Design
Reporting Guidelines
Provided For
Reporting Guideline
Acronym
Full Text if
Available
Explanation and
Elaboration
Reporting systematic
reviews in journal and
conference abstracts
Meta-analysis of individual
participant data
Economic evaluations
Economic evaluations of
health interventions
CHEERS
http://www.ispor.org/
taskforces/Economic
PubGuidelines.asp
PMID: 19622551.
Ann Intern Med. 2009;
151(4):264269, W64.
PMID: 19622511 (83).
J Clin Epidemiol. 2009;
62(10):10061012. PMID:
19631508 (84).
Open Med. 2009; 3(3);
123130
PLoS Med. 2013; 10(4):
e1001419. PMID:
23585737 (85).
CRONIN ET AL
1096
TABLE 5. (continued) Reporting Guidelines by Research Study Design, Acronym, Web site URL, and Bibliographic Reference
Qualitative research
interviews and focus
groups
COREQ
Observational studies
For completeness,
transparency and data
analysis in case reports
and data from the point of
care.
CARE
GRRAS
Qualitative research,
systematic reviews/
meta-analyses/HTA
Synthesis of qualitative
research
ENTREQ
Qualitative research
Qualitative research
interviews and focus
groups
COREQ
Mixed-methods studies
GRAMMS
http://www.care-statement.
org/
http://intqhc.oxfordjournals.
org/content/19/6/349.
long
Full text
L, Brorson S, Donner
Kottner J, Audige
bjartsson A,
A, Gajeweski BJ, Hro
Robersts C, Shoukri M, Streiner DL.
Guidelines for reporting reliability and
agreement studies (GRRAS) were
proposed.
Tong A, Flemming K, McInnes E, Oliver
S, Craig J. Enhancing transparency in
reporting the synthesis of qualitative
research: ENTREQ.
Tong A, Sainsbury P, Craig J.
Consolidated criteria for reporting
qualitative research (COREQ): a 32item checklist for interviews and focus
groups.
OCathain A, Murphy E, Nicholl J. The
quality of mixed methods studies in
health services research.
1097
Narrative in reports of
medical research
Research
Study Design
Reporting Guidelines
Provided For
Reporting Guideline
Acronym
Quality improvement
studies
Quality improvement in
health care
SQUIRE
Health informatics
STARE-HI
Full Text if
Available
CRONIN ET AL
1098
TABLE 5. (continued) Reporting Guidelines by Research Study Design, Acronym, Web site URL, and Bibliographic Reference
CARE, case reports; CHEERS, consolidated health economic evaluation reporting standards; CONSORT, consolidated standards of reporting trials; COREQ, consolidated criteria for reporting qualitative research; ENTREQ, enhancing transparency in reporting the synthesis of qualitative research; GRAMMS, good reporting of a mixed-methods study; GRRAS, guidelines for reporting reliability and agreement studies; HTA, health technology assessment; PRISMA, preferred reporting items for systematic reviews and meta-analyses; SPIRIT, standard protocol items:
recommendations for interventional trials; SQUIRE, standards for quality improvement reporting excellence; STARD, standards for reporting of diagnostic accuracy; STARE-HI, statement on
reporting of evaluation studies in health informatics.
Therapeutic Studies
specific issues that introduce bias to results reporting. This includes challenges because of impossible or partial blinding,
clustering, the experience of providers and centers, and
both patient and provider willingness to undergo randomization (45). Thus, the CONSORT Statement has been specifically modified to guide reporting of nonpharmacologic
treatments (46). In addition, modifications have been made
to promote systematic reporting for cohort and comparative
study designs, using the principles from the CONSORT
Statement (45,47).
The most critical modifications to the CONSORT
Statement for reporting interventional and therapeutic study
results relate to treatment and provider details. Precise descriptions of the experimental treatment and comparator should be
reported (48). In addition, descriptions of how, why, and
when treatment is modified help to demonstrate sufficient
separation of study arms, as study arm crossover may limit
the conclusions that can be drawn from interventional trials
(49). Although rarely included, it is recommended that trials
describe in detail the volume of experience of providers, as patient outcomes are directly related to experience (48,50).
Correlations between patients could be introduced on the
basis of undocumented similarities in process for instance by
a single provider or practice location. Clustering is the instance
in which the subjects in one trial arm are more like each other
than the subjects in another arm, thus reducing statistical power and introducing a challenge to interpreting study results.
For example, if therapy A is only offered in a homogenous suburban community and therapy B is offered only in a highdensity urban center, the subjects who undergo therapy A
will tend to be like each other but unlike those undergoing
therapy B. If the outcome is better in the population undergoing therapy A, fewer patients will be required to demonstrate a
statistically significant benefit of therapy A compared to therapy B. However, the benefit is due to the similarity of patients
by group, rather than the therapy itself. To overcome this clustering, it is necessary to recruit more patients and explicitly account for the similarities in each group. This accounting
should be described in the sample size calculation and statistical
reporting. Alternatively, if both therapy A and therapy B were
offered in either center and subjects were randomly assigned to
either therapy, sample size could be reduced.
The main CONSORT Statement is based on the standard two-group parallel design. However, there are several
different types of randomized trials, some of which have
different designs, interventions, and data. To help improve
the reporting of these trials, the CONSORT Group has
been involved in extending and modifying the main
CONSORT Statement for application in these various areas
including design extensions such as cluster trials, noninferiority and equivalence trials and pragmatic trials; intervention extensions for nonpharmacological treatment interventions; and
data extensions for patient-reported outcomes and harms.
Future directions for CONSORT are new CONSORT extensions and update the various CONSORT extensions to
reflect the 2010 checklist. For additional details, it is
1099
CRONIN ET AL
recommended that the reader review the Extended CONSORT statement, which may be found on the EQUATOR
network (1,51).
Meta-analyses
Types of medical education studies include curricular innovations which may follow the Kern six-step process (including
problem identification, targeted needs assessment, goals and
objectives, deciding educational strategies, implementation,
and evaluation); consensus conference proceedings, identifying and addressing knowledge gaps which may use a formal
process to achieve consensus such as the Delphi method; qualitative research studies; quantitative research studies; and
mixed-methods research studies (65,66). Curricular
innovations can be subjective, assessing learner satisfaction
or self-reported confidence or objective assessing knowledge,
skills, attitudes, behaviors, or performance (65). Educational
studies can be descriptive, such as case reports/case series,
correlational (ecologic) studies, or cross-sectional studies.
They can also be analytical, such as casecontrol studies,
cohort/prospective studies, or RCTs (67). Medical education
study designs include true experimental designs, such as prospective cohort studies which have a pretest or post-test with
control group design; the Solomon four-group design, which
has two intervention and two control groups, half of each
group taking the pretest and all taking the post-test; and
the post-test only with control group design. Quasiexperimental designs include time series design (repeated
1101
CRONIN ET AL
27. Bossuyt PM, Reitsma JB, Bruns DE, et al. Towards complete and accurate reporting of studies of diagnostic accuracy: The STARD Initiative.
Ann Intern Med 2003; 138(1):4044.
28. Pai M, Sharma S. Better reporting of studies of diagnostic accuracy. Indian J Med Microbiol 2005; 23(4):210213.
29. Liberati A, Altman DG, Tetzlaff J, et al. The PRISMA statement for reporting systematic reviews and meta-analyses of studies that evaluate health
care interventions: explanation and elaboration. PLoS Med 2009; 6(7):
e1000100.
30. Husereau D, Drummond M, Petrou S, et al. Consolidated Health Economic Evaluation Reporting Standards (CHEERS) statement. Eur J
Health Econ 2013; 14(3):367372.
31. Husereau D, Drummond M, Petrou S, et al. Consolidated Health Economic Evaluation Reporting Standards (CHEERS) statement. Value
Health 2013; 16(2):e1e5.
32. Husereau D, Drummond M, Petrou S, et al. Consolidated Health Economic Evaluation Reporting Standards (CHEERS) statement. Clin Ther
2013; 35(4):356363.
33. Husereau D, Drummond M, Petrou S, et al. Consolidated Health Economic Evaluation Reporting Standards (CHEERS) statement. Cost Eff
Resour Alloc 2013; 11(1):6.
34. Husereau D, Drummond M, Petrou S, et al. Consolidated Health Economic Evaluation Reporting Standards (CHEERS) statement. BMC Med
2013; 11:80.
35. Husereau D, Drummond M, Petrou S, et al. Consolidated Health Economic Evaluation Reporting Standards (CHEERS) statement. BMJ
2013; 346:f1049.
36. Husereau D, Drummond M, Petrou S, et al. Consolidated Health Economic Evaluation Reporting Standards (CHEERS) statement. Pharmacoeconomics 2013; 31(5):361367.
37. Husereau D, Drummond M, Petrou S, et al. Consolidated Health Economic Evaluation Reporting Standards (CHEERS) statement. J Med
Econ 2013; 16(6):713719.
38. Husereau D, Drummond M, Petrou S, et al. Consolidated Health Economic Evaluation Reporting Standards (CHEERS) statement. Int J Technol Assess Health Care 2013; 29(2):117122.
39. Husereau D, Drummond M, Petrou S, et al. Consolidated Health Economic Evaluation Reporting Standards (CHEERS) statement. BJOG
2013; 120(6):765770.
40. Stein PD, Fowler SE, Goodman LR, et al. Multidetector computed tomography for acute pulmonary embolism. N Engl J Med 2006; 354(22):
23172327.
41. The STARD website. http://www.stard-statement.org/. Accessed
February 7, 2014.
42. Johnston KC, Holloway RG. There is nothing staid about STARD: progress in the reporting of diagnostic accuracy studies. Neurology 2006;
67(5):740741.
43. The CONSORT website. http://www.consort-statement.org/. Accessed
February 7, 2014.
44. Moher D, Jones A, Lepage L, et al. Use of the CONSORT statement and
quality of reports of randomized trials: a comparative before-and-after
evaluation. JAMA 2001; 285(15):19921995.
45. Reeves BC, Gaus W. Guidelines for reporting non-randomised studies.
Forsch Komplementarmed Klass Naturheilkd 2004; 11(Suppl 1):4652.
46. Boutron I, Moher D, Altman DG, et al. Extending the CONSORT statement to randomized trials of nonpharmacologic treatment: explanation
and elaboration. Ann Intern Med 2008; 148(4):295309.
47. Reeves BC. A framework for classifying study designs to evaluate health
care interventions. Forsch Komplementarmed Klass Naturheilkd 2004;
11(Suppl 1):1317.
48. Salem R, Lewandowski RJ, Gates VL, et al. Research reporting standards
for radioembolization of hepatic malignancies. J Vasc Interv Radiol 2011;
22(3):265278.
49. Kallmes DF, Comstock BA, Heagerty PJ, et al. A randomized trial of vertebroplasty for osteoporotic spinal fractures. N Engl J Med 2009; 361(6):
569579.
50. Jacquier I, Boutron I, Moher D, et al. The reporting of randomized clinical
trials using a surgical intervention is in need of immediate improvement: a
systematic review. Ann Surg 2006; 244(5):677683.
51. Campbell MK, Piaggio G, Elbourne DR, et al. Consort 2010 statement:
extension to cluster randomised trials. BMJ 2012; 345:e5661.
52. Davey J, Turner RM, Clarke MJ, et al. Characteristics of meta-analyses
and their component studies in the Cochrane database of systematic
1103
CRONIN ET AL
53.
54.
55.
56.
57.
58.
59.
60.
61.
62.
63.
64.
65.
66.
67.
68.
69.
70.
71.
72.
73.
74.
75.
76.
1104
APPENDIX TABLE 1. STARD checklist for reporting of studies of diagnostic accuracy (1328).
Section and Topic
Item#
Title/Abstract/Keywords
Introduction
Methods
Participants
3
4
6
Test methods
7
8
9
10
11
Statistical methods
12
13
Results
Participants
14
15
16
Test results
17
18
19
Estimates
20
21
22
23
Discussion
24
25
On Page#
Identify the article as a study of diagnostic accuracy (recommend MeSH heading
sensitivity and specificity).
State the research questions or study aims, such as estimating diagnostic accuracy or
comparing accuracy between tests or across participant groups.
The study population: The inclusion and exclusion criteria, setting and locations where
data were collected.
Participant recruitment: Was recruitment based on presenting symptoms, results from
previous tests, or the fact that the participants had received the index tests or the
reference standard?
Participant sampling: Was the study population a consecutive series of participants
defined by the selection criteria in item 3 and 4? If not, specify how participants were
further selected.
Data collection: Was data collection planned before the index test and reference standard
were performed (prospective study) or after (retrospective study)?
The reference standard and its rationale.
Technical specifications of material and methods involved including how and when
measurements were taken, and/or cite references for index tests and reference
standard.
Definition of and rationale for the units, cut-offs and/or categories of the results of the
index tests and the reference standard.
The number, training and expertise of the persons executing and reading the index tests
and the reference standard.
Whether or not the readers of the index tests and reference standard were blind (masked)
to the results of the other test and describe any other clinical information available to the
readers.
Methods for calculating or comparing measures of diagnostic accuracy, and the statistical
methods used to quantify uncertainty (e.g. 95% confidence intervals).
Methods for calculating test reproducibility, if done.
When study was performed, including beginning and end dates of recruitment.
Clinical and demographic characteristics of the study population (at least information on
age, gender, spectrum of presenting symptoms).
The number of participants satisfying the criteria for inclusion who did or did not undergo
the index tests and/or the reference standard; describe why participants failed to
undergo either test (a flow diagram is strongly recommended).
Time-interval between the index tests and the reference standard, and any treatment
administered in between.
Distribution of severity of disease (define criteria) in those with the target condition; other
diagnoses in participants without the target condition.
A cross tabulation of the results of the index tests (including indeterminate and missing
results) by the results of the reference standard; for continuous results, the distribution
of the test results by the results of the reference standard.
Any adverse events from performing the index tests or the reference standard.
Estimates of diagnostic accuracy and measures of statistical uncertainty (e.g. 95%
confidence intervals).
How indeterminate results, missing data and outliers of the index tests were handled.
Estimates of variability of diagnostic accuracy between subgroups of participants,
readers or centers, if done.
Estimates of test reproducibility, if done.
Discuss the clinical applicability of the study findings.
1105
CRONIN ET AL
1106
Section/Topic
Item
No
Reported on
Page No
Checklist Item
Results
Participant flow (a diagram
is strongly recommended)
8a
8b
9
10
11a
11b
12a
12b
13a
Baseline data
Numbers analysed
13b
14a
14b
15
16
17a
Recruitment
1107
17b
Statistical methods
6a
6b
7a
7b
APPENDIX TABLE 2. CONSORT 2010 checklist of information to include when reporting a randomized trial (212).
Section/Topic
Ancillary analyses
Harms
Discussion
Limitations
Generalisability
Interpretation
Other information
Registration
Protocol
Funding
Item
No
18
Checklist Item
19
Results of any other analyses performed, including subgroup analyses and adjusted analyses, distinguishing pre-specified from
exploratory
All important harms or unintended effects in each group (for specific guidance see CONSORT for harms [28])
20
21
22
Trial limitations, addressing sources of potential bias, imprecision, and, if relevant, multiplicity of analyses
Generalisability (external validity, applicability) of the trial findings
Interpretation consistent with results, balancing benefits and harms, and considering other relevant evidence
23
24
25
Reported on
Page No
CRONIN ET AL
1108
APPENDIX TABLE 2. (continued) CONSORT 2010 checklist of information to include when reporting a randomized trial (212).
1109
Section/topic
Checklist Item
Title
1
Provide a structured summary including, as applicable: background; objectives; data sources; study eligibility criteria,
participants, and interventions; study appraisal and synthesis methods; results; limitations; conclusions and implications
of key findings; systematic review registration number.
Introduction
Rationale
Objectives
3
4
Describe the rationale for the review in the context of what is already known.
Provide an explicit statement of questions being addressed with reference to participants, interventions, comparisons,
outcomes, and study design (PICOS).
Methods
Protocol and registration
Indicate if a review protocol exists, if and where it can be accessed (e.g., Web address), and, if available, provide registration
information including registration number.
Specify study characteristics (e.g., PICOS, length of follow-up) and report characteristics (e.g., years considered, language,
publication status) used as criteria for eligibility, giving rationale.
Describe all information sources (e.g., databases with dates of coverage, contact with study authors to identify additional
studies) in the search and date last searched.
Present full electronic search strategy for at least one database, including any limits used, such that it could be repeated.
State the process for selecting studies (i.e., screening, eligibility, included in systematic review, and, if applicable, included
in the meta-analysis).
Describe method of data extraction from reports (e.g., piloted forms, independently, in duplicate) and any processes for
obtaining and confirming data from investigators.
List and define all variables for which data were sought (e.g., PICOS, funding sources) and any assumptions and
simplifications made.
Describe methods used for assessing risk of bias of individual studies (including specification of whether this was done at
the study or outcome level), and how this information is to be used in any data synthesis.
State the principal summary measures (e.g., risk ratio, difference in means).
Describe the methods of handling data and combining results of studies, if done, including measures of consistency (e.g., I2)
for each meta-analysis.
Specify any assessment of risk of bias that may affect the cumulative evidence (e.g., publication bias, selective reporting
within studies).
Describe methods of additional analyses (e.g., sensitivity or subgroup analyses, meta-regression), if done, indicating which
were pre-specified.
Eligibility criteria
Information sources
Search
Study selection
8
9
10
Data items
11
12
Summary measures
Synthesis of results
13
14
15
Additional analyses
16
Results
Study selection
17
Study characteristics
18
19
Give numbers of studies screened, assessed for eligibility, and included in the review, with reasons for exclusions at each
stage, ideally with a flow diagram.
For each study, present characteristics for which data were extracted (e.g., study size, PICOS, follow-up period) and
provide the citations.
Present data on risk of bias of each study and, if available, any outcome level assessment (see item 12).
Title
Abstract
Structured summary
Reported on
Page#
CRONIN ET AL
1110
APPENDIX TABLE 3. The PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) checklist (29).
Synthesis of results
Risk of bias across studies
Additional analysis
Discussion
Summary of evidence
20
21
22
23
24
Limitations
25
Conclusions
Funding
Funding
26
27
For all outcomes considered (benefits or harms), present, for each study: (a) simple summary data for each intervention
group (b) effect estimates and confidence intervals, ideally with a forest plot.
Present results of each meta-analysis done, including confidence intervals and measures of consistency.
Present results of any assessment of risk of bias across studies (see Item 15).
Give results of additional analyses, if done (e.g., sensitivity or subgroup analyses, meta-regression [see Item 16]).
Summarize the main findings including the strength of evidence for each main outcome; consider their relevance to key
groups (e.g., healthcare providers, users, and policy makers).
Discuss limitations at study and outcome level (e.g., risk of bias), and at review-level (e.g., incomplete retrieval of identified
research, reporting bias).
Provide a general interpretation of the results in the context of other evidence, and implications for future research.
Describe sources of funding for the systematic review and other support (e.g., supply of data); role of funders for the
systematic review.
1111
CRONIN ET AL
1112
Section/Item
Title and abstract
Title
Abstract
Introduction
Background and
objectives
Methods
Target population and
subgroups
Setting and location
Study perspective
Comparators
Time horizon
Discount rate
Choice of health
outcomes
Measurement of
effectiveness
Item
No.
1
2
Describe characteristics of the base case population and subgroups analysed, including why they were chosen.
11a
11b
12
13a
13b
14
Choice of model
15
Assumptions
16
State relevant aspects of the system(s) in which the decision(s) need(s) to be made.
Describe the perspective of the study and relate this to the costs being evaluated.
Describe the interventions or strategies being compared and state why they were chosen.
State the time horizon(s) over which costs and consequences are being evaluated and say why appropriate.
Report the choice of discount rate(s) used for costs and outcomes and say why appropriate.
Describe what outcomes were used as the measure(s) of benefit in the evaluation and their relevance for the type of analysis
performed.
Single study-based estimates: Describe fully the design features of the single effectiveness study and why the single study
was a sufficient source of clinical effectiveness data.
Synthesis-based estimates: Describe fully the methods used for identification of included studies and synthesis of clinical
effectiveness data.
If applicable, describe the population and methods used to elicit preferences for outcomes.
Single study-based economic evaluation: Describe approaches used to estimate resource use associated with the
alternative interventions. Describe primary or secondary research methods for valuing each resource item in terms of its
unit cost. Describe any adjustments made to approximate to opportunity costs.
Model-based economic evaluation: Describe approaches and data sources used to estimate resource use associated with
model health states. Describe primary or secondary research methods for valuing each resource item in terms of its unit
cost. Describe any adjustments made to approximate to opportunity costs.
Report the dates of the estimated resource quantities and unit costs. Describe methods for adjusting estimated unit costs to
the year of reported costs if necessary. Describe methods for converting costs into a common currency base and the
exchange rate.
Describe and give reasons for the specific type of decision-analytical model used. Providing a figure to show model
structure is strongly recommended.
Describe all structural or other assumptions underpinning the decision-analytical model.
1113
Measurement and
valuation of preference
based outcomes
Estimating resources
and costs
Identify the study as an economic evaluation or use more specific terms such as cost-effectiveness analysis, and
describe the interventions compared.
Provide a structured summary of objectives, perspective, setting, methods (including study design and inputs), results
(including base case and uncertainty analyses), and conclusions.
5
6
7
8
9
10
Reported on Page
No./Line No.
Recommendation
APPENDIX TABLE 4. Consolidated Health Economic Evaluation Reporting Standards (CHEERS) checklist. Items to include when reporting economic evaluations of health
interventions (3039).
Section/Item
Analytical methods
Results
Study parameters
Item
No.
17
Describe all analytical methods supporting the evaluation. This could include methods for dealing with skewed, missing, or
censored data; extrapolation methods; methods for pooling data; approaches to validate or make adjustments (such as
half cycle corrections) to a model; and methods for handling population heterogeneity and uncertainty.
18
Report the values, ranges, references, and, if used, probability distributions for all parameters. Report reasons or sources
for distributions used to represent uncertainty where appropriate.
Providing a table to show the input values is strongly recommended.
For each intervention, report mean values for the main categories of estimated costs and outcomes of interest, as well as
mean differences between the comparator groups. If applicable, report incremental cost-effectiveness ratios.
Single study-based economic evaluation: Describe the effects of sampling uncertainty for the estimated incremental cost
and incremental effectiveness parameters, together with the impact of methodological assumptions (such as discount
rate, study perspective).
Model-based economic evaluation: Describe the effects on the results of uncertainty for all input parameters, and
uncertainty related to the structure of the model and assumptions.
If applicable, report differences in costs, outcomes, or costeffectiveness that can be explained by variations between
subgroups of patients with different baseline characteristics or other observed variability in effects that are not reducible
by more information.
19
20a
20b
Characterising
heterogeneity
Conflicts of interest
21
22
Summarise key study findings and describe how they support the conclusions reached. Discuss limitations and the
generalisability of the findings and how the findings fit with current knowledge.
23
Describe how the study was funded and the role of the funder in the identification, design, conduct, and reporting of the
analysis. Describe other non-monetary sources of support.
Describe any potential for conflict of interest of study contributors in accordance with journal policy. In the absence of a
journal policy, we recommend authors comply with International Committee of Medical Journal Editors
recommendations.
24
Reported on Page
No./Line No.
Discussion
Study findings,
limitations,
generalisability, and
current knowledge
Other
Source of funding
Recommendation
CRONIN ET AL
1114
APPENDIX TABLE 4. (continued) Consolidated Health Economic Evaluation Reporting Standards (CHEERS) checklist. Items to include when reporting economic evaluations of
health interventions (3039).
APPENDIX TABLE 5. Consolidated criteria for reporting qualitative studies (COREQ): 32-item checklist (73).
No.
Item
Interviewer characteristics
Guide Questions/Description
What methodological orientation was stated to underpin the study? e.g. grounded
theory, discourse analysis, ethnography, phenomenology, content analysis
How were participants selected? e.g. purposive, convenience, consecutive,
snowball
How were participants approached? e.g. face-to-face, telephone, mail, email
How many participants were in the study?
How many people refused to participate or dropped out? Reasons?
Where was the data collected? e.g. home, clinic, workplace
Was anyone else present besides the participants and researchers?
What are the important characteristics of the sample? e.g. demographic data, date
Were questions, prompts, guides provided by the authors? Was it pilot tested?
Were repeat interviews carried out? If yes, how many?
Did the research use audio or visual recording to collect the data?
Were field notes made during and/or after the interview or focus group?
What was the duration of the interviews or focus group?
Was data saturation discussed?
Were transcripts returned to participants for comment and/or correction?
1115
CRONIN ET AL
APPENDIX TABLE 6. Enhancing transparency in reporting the synthesis of qualitative research: the ENTREQ statement (74)
No.
1
2
3
4
5
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
1116
Item
Aim
Synthesis methodology