Professional Documents
Culture Documents
ORIGINAL ARTICLE
Abstract
Objectives: The objective of the study is to identify items and domains applicable for the quality assessment of prevalence studies.
Study Design and Setting: We searched databases and the gray literature to identify tools or guides about the quality assessment of
prevalence studies. After study selection, we abstracted questions applicable for prevalence studies and classified into at least one of the
following domains: ‘population and setting’, ‘condition measurement’, ‘statistics’, and ‘other’. PROSPERO registration:CRD42018088437.
Results: We included 30 tools: eight (26.7%) specifically designed to appraise prevalence studies and 22 (73.3%) adaptable for this
purpose. We identified 12 unique items in the domain ‘‘population and setting’’, 16 in the domain ‘‘condition measurement’’, and 14 in
the domain ‘‘statistics’’. Of those, 25 (59.5%) were identified in the eight specific tools. Regarding the domain ‘‘other’’, we identified
77 unique items, mainly related to manuscript writing and reporting (n 5 48, 62.3%); of those, 24 (31.2%) were identified in the eight
specific tools and 53 (68.8%) in the additional 22 nonspecific tools.
Conclusion: We provide a comprehensive set of items classified by domains that can guide the appraisal of prevalence studies, con-
duction of primary prevalence studies, and update or development of tools to evaluate prevalence studies. Ó 2020 The Authors. Published
by Elsevier Inc. This is an open access article under the CC BY-NC-ND license (http://creativecommons.org/licenses/by-nc-nd/4.0/).
https://doi.org/10.1016/j.jclinepi.2020.06.039
0895-4356/Ó 2020 The Authors. Published by Elsevier Inc. This is an open access article under the CC BY-NC-ND license (http://creativecommons.org/
licenses/by-nc-nd/4.0/).
60 C.B. Migliavaca et al. / Journal of Clinical Epidemiology 127 (2020) 59e68
Yes No
Yes No
Fig. 1. Flowchart of classification of questions/statements. *In some instruments, the guidance about which questions/statements were applicable
for prevalence studies was based on study aspects such as intervention or the comparison group.
reporting or study methods), it was included under the clas- were solved by consensus or arbitrated by a third reviewer
sification ‘‘other’’. (V.C. or M.F.).
As described, we included not only tools specifically de-
signed to assess prevalence studies but also tools that could
be adapted for this purpose. Thus, not all questions from 3. Results
nonspecific tools were applicable for prevalence studies (such
as questions assessing the comparability among groups, or the 3.1. Study selection
description of intervention), and we only categorized the Our search resulted in 1,690 unique references. After se-
applicable ones. If the instrument provided guidance about lection of titles and abstracts, we assessed 105 full texts for
which questions should be used to assess prevalence studies, eligibility. Finally, we included in the review 30 tools
we followed these instructions. If not, before classyifing the [11e41]. Fig. 2 presents the flowchart of study selection.
questions into items and domains, we evaluated if they were The list of full texts excluded with reason is available in
applicable for prevalence studies or not. If classified as appli- Additional file 2.
cable, the question was categorized into items and domains as
previously described. Questions classified as not applicable
were not further evaluated (Fig. 1). 3.2. Tools specific for prevalence studies
The process of selection and classification of questions/ Out of the 30 tools, eight (26.7%) were specifically de-
statements into items and domains was conducted by two signed to appraise prevalence studies [11e19]. Table 1 sum-
reviewers independently (C.B.M. and C.S.). Discrepancies marizes the main characteristics of these tools. Among these
Duplicates removed
(n = 1.235)
Records screened
(n = 1.689)
Records excluded
(n = 1.590)
Full-text assessed for elegilibity
(n = 99)
Full text articles excluded
(n = 67)
Does not describe new tool (n = 36)
Full text included in qualitative Tool not applicable for prevalence studies (n = 28)
Other (n = 3)
synthesis
(n = 32)
Number of tools = 30
(Continued )
C.B. Migliavaca et al. / Journal of Clinical Epidemiology 127 (2020) 59e68 63
Table 1. Continued
Instrument (name of the
instrument or author, Context of development Summary and reporting of
year) (clinical condition) Process of development Structure results
The Joanna Briggs Any clinical condition. Systematic search for Nine questions with four Question for overall
Institute Prevalence instruments to assess standard answer appraisal with three
Critical Appraisal prevalence studies; options ( yes/no/ answer options
Tool, 2014 review and selection unclear/not (include/exclude/seek
[[18],[19]] of applicable criteria; applicable).a further info), based
development of the on rater’s judgment.
draft tool; pilot tests
with professionals to
assess face validity,
applicability,
acceptability,
timeliness, and ease
of use; the final
version of the tool.
tools, seven (87.5%) were new tools [11e13,15e19], and quality of prevalence studies, six related to ‘‘population
one (12.5%) was an adaptation of an existing instrument and setting’’, nine related to ‘‘condition measurement’’,
[14]. Four tools (50.0%) were developed to assess studies and 10 related to ‘‘statistics’’. In addition, 24 items were
of prevalence of any clinical condition [13e15,18,19], classified as ‘‘other’’, mainly related to reporting (Table 2).
whereas the other four tools (50.0%) were developed to
appraise prevalence studies of specific medical fields
3.3. Tools adapted for prevalence studies
[11,12,16,17]; however, with some adaptations, they could
all be applied to any clinical condition. The process of devel- Among the 30 included tools, 22 (73.3%) were not spe-
opment of all instruments included search and review of rele- cific for prevalence studies. The main characteristics of
vant criteria, piloting test(s), and adjustments for the final these tools and the items and domains assessed by them
version. The median number of questions in the tools was are presented in Additional file 3 and 4, respectively. These
10, ranging from seven to 32. Four tools (50.0%) were scales, tools provided six unique additional items for the domain
with numeric results [11,13,15,17], and four (50.0%) were ‘‘population and setting’’, seven items related to ‘‘condition
descriptive checklists [12,14,16,18,19]. Among the scales, measurement’’, and four items related to ‘‘statistics’’.
only one suggested cutoff values to define the overall quality Moreover, these tools provided 53 items classified under
of the study [13]; among the checklists, two had an overall the domain ‘‘other’’ (Additional file 4).
appraisal question, but they were answered based on rater’s
judgment, without guidance of how to consider the previous
questions to define a summary assessment [14,18,19]. 3.4. Items
Regarding the domains assessed, seven tools (87.5%) We identified 710 questions/statements from the
covered all key domains [12e19]; however, there was vari- included tools that were compiled into 119 different items.
ability regarding the items assessed by each tool. Table 2 We identified 42 unique items classified under the do-
describes the items of each tool, classified by domains. In mains ‘‘population and setting’’ (12 items), ‘‘condition
the domain ‘‘population and setting’’, the main items as- measurement’’ (16 items), and ‘‘statistics’’ (14 items); of
sessed were ‘‘appropriate sampling’’ (seven tools, 87.5%), those, 25 (59.5%) were identified in the eight specific tools
‘‘appropriate response rate’’ (five tools, 62.5%), and and 17 (40.5%) were identified in the additional 22 nonspe-
‘‘representative sample’’ (four tools, 50.0%). In the domain cific tools. Table 3 summarized the items assessed in each
‘‘condition measurement’’, the main items assessed were domain among all tools.
‘‘valid measurement of condition’’ (six tools, 75.0%), In the domain ‘‘other’’, we identified 77 unique items; of
‘‘standard measurement of condition’’ (six tools, 75.0%), those, 24 (31.2%) were identified in the eight specific tools
and ‘‘reliable measurement of condition’’ (five tools, and 53 (68.8%) were identified in the additional 22 nonspe-
62.5%). In the domain ‘‘statistics’’, the main items assessed cific tools. We classified the items in the domain ‘‘other’’
were ‘‘precision of estimate’’ (six tools, 75.0%), ‘‘data into three categories: ‘‘manuscript writing and reporting’’,
analysis considering sampling’’ (three tools, 37.5%), ‘‘study protocol and methods’’, and ‘‘nonclassified’’. These
‘‘appropriate sample size’’ (two tools, 25.0%), and ‘‘sub- categories were defined after data extraction, based on our
group analysis’’ (two tools, 25.0%). Overall, these tools findings. Table 4 presents the items classified as ‘‘other’’
provided 25 unique items related to the assessment of stratified by these categories.
64 C.B. Migliavaca et al. / Journal of Clinical Epidemiology 127 (2020) 59e68
Table 2. Items assessed by tools specifically designed to appraise prevalence studies, classified by domainsa
4. Discussion Not all domains were covered by all tools, and even
when they were covered, they were not always properly as-
In this systematic review, we identified, summarized, sessed. Some tools did not consider important aspects in-
and compared 30 instruments used for the quality assess-
side each domain, such as representativeness of sample,
ment of prevalence studies. Our results, similar to what
estimation of sample size, and appropriate measurement
was found in other reviews, show that there is great vari-
of condition, and there was an overlap among questions
ability among tools and there is no consensus about which
in the same instrument, which may lead to penalization
domains should be assessed in prevalence studies [42,43].
of the same study for the same reason more than once.
We classified all questions or statements into items and do-
Moreover, many instruments assessed not only risk of bias
mains, creating a comprehensive set of 119 items useful for
but also reporting and manuscript writing. It is important to
the assessment of prevalence studies.
C.B. Migliavaca et al. / Journal of Clinical Epidemiology 127 (2020) 59e68 65
Table 3. Unique items identified among all included tools and classified into a key domain
Population and setting (n [ 12) Condition measurement (n [ 16) Statistics (n [ 14)
Appropriate samplea (seven tools) Type of prevalence estimate (point or Sample size estimationa (10 tools)
Unbiased samplea (one tool) period) (one tool) Appropriate sample size (seven tools)
Representative sample (14 tools) Appropriate length of prevalence period Appropriate statistical analysis (13
Appropriate sample source (two tools) (two tools) tools)
Appropriate size of population source Appropriate definition of condition (two Appropriate numerator and denomina-
(one tool) tools) tor parameters (one tool)
Ethnic characteristics of population Appropriate measurement of condition Appropriate exclusion from analysis
source (one tool) (11 tools) (one tool)
Appropriate sampling (15 tools) Accurate measurement of conditiona Adjustment of estimates (one tool)
Random samplinga (one tool) (one tool) Data analysis considering sampling
Standard selection of participantsa (two Precise measurement of conditiona (three tools)
tools) (one tool) Data analysis considering the response
Participation rate of eligible personsa Quality control of measurement meth- rate (seven tools)
(one tool) odsa (two tools) Data analysis considering special fea-
Appropriate response rate (19 tools) Valid measurement of condition (21 tures (one tool)
Assessment of nonrespondersa (one tools) Missing data handlinga (one tool)
tool) Reliable measurement of condition (15 Random errora (three tools)
tools) Precision of estimate (11 tools)
Standard measurement of condition Subgroup analysis (five tools)
(10 tools) Data fishinga (two tools)
Unbiased measurement of condition
(three tools)
Reproducible measurement of condi-
tiona (two tools)
Assessment of disease severity and
frequency of symptoms (one tool)
Data collection performed by investi-
gators unrelated to patientsa (one tool)
Face validitya (one tool)
Selective outcome reportinga (one tool)
a
Items from nonspecific tools only.
distinguish between these two concepts as poor reporting is statements into items and domains are available in
not a reflection on the quality of a study or whether the re- Additional file 5.
sults from a study are at risk of bias. The main objective of this study was to map the literature
We conducted a broad search, using important databases to identify items used to assess the risk of bias of prevalence
and including alternative data sources. Our search was very studies, generating a comprehensive bank of items. Some of
sensitive to identify tools specifically designed for preva- the items identified are similar and there may be overlap
lence studies, but we probably have not included all instru- among them in terms of their broad definition. However,
ments that could be adapted for this purpose. This could be we do believe that subtle differences in terminology and
a limitation of our study; however, we believe our results nomenclature are important in the contextualization of the
are representative of the items and domains used to tools in which they are derived. This granular approach is
appraise prevalence studies because we probably achieved justified in this case as this will facilitate the future develop-
a saturation of items, and this work is the most comprehen- ment of a new risk of bias assessment tool. As an example of
sive overview of tools to assess prevalence studies to date. the importance of identifying subtle differences between
Another possible limitation of our review is that data similar items in tools, we attempt in the following to clarify
abstraction and classification of questions into items and the differences between valid, reliable, reproducible, and un-
domains required judgment, which can lead to different de- biased measurement of the condition (which some readers
cisions by different assessors. We tried to overcome this by may have considered synonymous terms previously):
conducting the classification independently by two re-
Valid measurement of the condition: the measurement
viewers with the assistance of third reviewers in case of dis-
of the condition is performed with methods that actu-
crepancies. In addition, our decision-making was informed
ally measures or detects what it is supposed to
by a protocol which reduced the chance of making ad hoc,
measure.
subjective decisions during the conduct of our review.
Reliable measurement of the condition: this is related
Moreover, to enhance transparency of the process, our
to the consistency of a measure. A highly reliable
judgments regarding the classification of questions and
measure produces similar results under similar
66 C.B. Migliavaca et al. / Journal of Clinical Epidemiology 127 (2020) 59e68
(Continued )
C.B. Migliavaca et al. / Journal of Clinical Epidemiology 127 (2020) 59e68 67
Table 4. Continued
Manuscript writing and reporting (n [ 48) Study protocol and methods (n [ 12) Nonclassified (n [ 17)
Comparison of results with the existing
literaturea (three tools)
Description of study conclusionsa (one
tool)
Appropriate conclusionsa (two tools)
Conclusion based on resultsa (10 tools)
Reporting of an additional information
sourcea (one tool)
Description of funding (two tools)
Reporting of conflict of interest (four
tools)
Clear referencesa (one tool)
Recommendations for future researcha
(two tools)
a
Items from nonspecific tools only.
conditions, so all things being equal, repeated testing Cinara Stein: Formal analysis, Investigation, Writing - re-
should produce similar results. view & editing. Ver^onica Colpani: Validation, Project
Reproducible measurement of the condition: the mea- administration, Writing - review & editing. Zachary
surement of the condition is performed using methods Munn: Methodology, Writing - review & editing. Maicon
capable of being reproduced at a different time or Falavigna: Conceptualization, Validation, Writing - review
place and by different people & editing.
Unbiased measurement of the condition: measure-
ment of the condition free of systematic errors that
could deviate the results from the truth.
Supplementary data
Therefore, even though the differences among items are
subtle, we do believe they are important to inform the Supplementary data to this article can be found online at
development of a new tool. https://doi.org/10.1016/j.jclinepi.2020.06.039.
It is not possible to strongly recommend a tool because
there is great variability in their content. A new tool, domain
based and with broader coverage and applicability is needed. References
However, among the currently available tools specific for [1] Fletcher R, Fletcher S, Fletcher GS. Clinical epidemiology: the es-
prevalence studies, the Joanna Briggs Institute Prevalence sentials. 5th ed. Philadelphia: Lippincott Williams & Wilkins; 2013.
Critical Appraisal Tool has a higher methodologic rigor [2] Harder T. Some notes on critical appraisal of prevalence studies:
and addresses what we consider the most important items comment on: "The development of a critical appraisal tool for use
in systematic reviews addressing questions of prevalence. Int J Health
related to the methodological quality of prevalence studies
Pol Manag 2014;3:289e90.
and may be considered the most appropriate tool [18,19]. [3] Wagner MB. Medindo a corr^eencia da doença: preval^encia ou in-
cid^encia? J Pediatr (Rio J 1998;74:157e62.
[4] Oxman AD, Schunemann HJ, Fretheim A. Improving the use of
research evidence in guideline development: 2. Priority setting.
5. Conclusions Health Res Pol Syst 2006;4:14.
[5] Rotily M, Roze S. What is the impact of disease prevalence upon
We have now identified a comprehensive set of items health technology assessment? Best Pract Res Clin Gastroenterol
and domains that is broader than any of the individual tools 2013;27:853e65.
included in this review. This data set can be now be used by [6] Borges Migliavaca C, Stein C, Colpani V, et al. How are systematic
those interested in the critical appraisal of prevalence reviews of prevalence conducted? A methodological study. BMC
Med Res Methodol 2020;20:96. https://doi.org/10.1186/s12874-020-
studies, by authors of prevalence studies, and to inform
00975-3.
the development or update of future tools to critically [7] Shamliyan T, Kane RL, Dickinson S. A systematic review of tools
appraise prevalence studies. used to assess the quality of observational studies that examine inci-
dence or prevalence and risk factors for diseases. J Clin Epidemiol
2010;63:1061e70.
[8] Sterne JA, Hernan MA, Reeves BC, Savovic J, Berkman ND,
CRediT authorship contribution statement Viswanathan M, et al. ROBINS-I: a tool for assessing risk of bias
in non-randomised studies of interventions. BMJ 2016;355:i4919.
Celina Borges Migliavaca: Formal analysis, Investiga- [9] Whiting PF, Rutjes AW, Westwood ME, Mallett S, Deeks JJ,
tion, Visualization, Data curation, Writing - original draft. Reitsma JB, et al. QUADAS-2: a revised tool for the quality
68 C.B. Migliavaca et al. / Journal of Clinical Epidemiology 127 (2020) 59e68
assessment of diagnostic accuracy studies. Ann Intern Med 2011;155: [27] Fowkes FG, Fulton PM. Critical appraisal of published research:
529e36. introductory guidelines. BMJ 1991;302:1136e40.
[10] Sterne JAC, Page MJ, Elbers RG, Blencowe NS, Boutron I, Cates CJ, [28] Gardner MJ, Machin D, Campbell MJ. Use of check lists in assessing
et al. RoB 2: a revised tool for assessing risk of bias in randomised the statistical content of medical studies. Br Med J (Clin Res Ed)
trials. BMJ 2019;366:l4898. 1986;292:810e2.
[11] Al-Jader LN, Newcombe RG, Hayes S, Murray A, Layzell J, [29] Glynn L. A critical appraisal tool for library and information
Harper PS. Developing a quality scoring system for epidemiological research. Libr Hi Tech 2006;24:387e99.
surveys of genetic disorders. Clin Genet 2002;62:230e4. [30] Kmet LS, Lee RC. Standard quality assessment criteria for evaluating
[12] Boyle MH. Guidelines for evaluating prevalence studies. Evid Based primary research papers from a variety of fieldsAHFMRHTA initia-
Ment Health 1998;1:37. tive20040213. HTA Initiative 2004:2.
[13] Giannakopoulos NN, Rammelsberg P, Eberhard L, Schmitter M. A [31] Law M, Stewart D, Pollock N, Letts L, Bosch J, Westmorland MIn:
new instrument for assessing the quality of studies on prevalence. Critical review formdquantitative studies, 20 1998:13.
Clin Oral Invest 2012;16:781e8. [32] Margetts B, Vorster H, Venter C. Evidence-based nutrition-review of
[14] Hoy D, Brooks P, Woolf A, Blyth F, March L, Bain C, et al. Assess- nutritional epidemiological studies. South Afr J Clin Nutr 2002;15:
ing risk of bias in prevalence studies: modification of an existing tool 68e73.
and evidence of interrater agreement. J Clin Epidemiol 2012;65: [33] Slim K, Nini E, Forestier D, Kwiatkowski F, Panis Y, Chipponi J.
934e9. Methodological index for non-randomized studies (minors): develop-
[15] Loney PL, Chambers LW, Bennett KJ, Roberts JG, Stratford PW. ment and validation of a new instrument. ANZ J Surg 2003;73:
Critical appraisal of the health research literature: prevalence or inci- 712e6.
dence of a health problem. Chronic Dis Can 1998;19:170e6. [34] Hong QN, Pluye P, Fabregues S, Bartlett G, Boardman F, Cargo M,
[16] Shamliyan TA, Kane RL, Ansari MT, Raman G, Berkman ND, et al. Mixed methods appraisal tool (MMAT), version 2018. Canada:
Grant M, et al. Development quality criteria to evaluate nonthera- IC Canadian Intellectual Property Office, Industry Canada; 2018.
peutic studies of incidence, prevalence, or risk factors of chronic dis- [35] Wells G, Shea B, O’Connell D, Peterson J, Welch V, Losos M, et al.
eases: pilot study of new checklists. J Clin Epidemiol 2011;64: Newcastle-Ottawa quality assessment scale cohort studies 2014.
637e57. Available at http://www.ohri.ca/programs/clinical_epidemiology/
[17] Silva LC, Ordunez P, Paz Rodriguez M, Robles S. A tool for assess- oxford.asp. Accessed August 1, 2020.
ing the usefulness of prevalence studies done for surveillance pur- [36] NIH. Quality Assessment tool for observational cohort and cross-
poses: the example of hypertension. Rev Panam Salud Publica sectional studies 2018. Available at https://www.nhlbi.nih.gov/health-
2001;10:152e60. topics/study-quality-assessment-tools. Accessed August 1, 2020.
[18] Munn Z, Moola S, Riitano D, Lisy K. The development of a critical [37] Public Health Wales Observatory. Critical appraisal checklist: cross
appraisal tool for use in systematic reviews addressing questions of sectional study 2004. Available at http://www2.nphs.wales.nhs.uk:
prevalence. Int J Health Pol Manag 2014;3:123e8. 8080/PubHObservatoryProjDocs.nsf/($All)/
[19] Munn Z, Moola S, Lisy K, Riitano D, Tufanaru C. Methodological E7B0C80995DC1BA380257DB80037C699/$File/Cross%
guidance for systematic reviews of observational epidemiological 20sectional%20study%20checklist.docx?OpenElement. Accessed
studies reporting prevalence and cumulative incidence data. Int J August 1, 2020.
Evid Based Healthc 2015;13:147e53. [38] Wong WCW, Cheung CSK, Hart GJ. Development of a quality
[20] Academy of Nutrition and Dietetics. Evidence analysis manual: steps assessment tool for systematic reviews of observational studies (QAT-
in the academy evidence analysis process. Chicago: Academy of SO) of HIV prevalence in men having sex with men and associated
Nutrition and Dietetics; 2016:106. risk behaviours. Emerging Themes Epidemiol 2008;5:23.
[21] Avis M. Reading research critically. II. An introduction to appraisal: [39] Viswanathan M, Berkman ND. Development of the RTI item bank on
assessing the evidence. J Clin Nurs 1994;3:271e7. risk of bias and precision of observational studies. J Clin Epidemiol
[22] Downes MJ, Brennan ML, Williams HC, Dean RS. Development of a 2012;65:163e78.
critical appraisal tool to assess the quality of cross-sectional studies [40] Viswanathan M, Berkman ND, Dryden DM, Hartling L. ahrq
(AXIS). BMJ Open 2016;6:7. methods for effective health care. assessing risk of bias and con-
[23] Berra S, Elorza-Ricart JM, Estrada MD, Sanchez E. [A tool (cor- founding in observational studies of interventions or exposures:
rected) for the critical appraisal of epidemiological cross-sectional further development of the RTI item bank. Rockville, MD: Agency
studies]. Gac Sanit 2008;22:492e7. for Healthcare Research and Quality (US); 2013.
[24] The University of Manchester. Centre for occupational and environ- [41] Specialist Unit for Review Evidence (SURE). Questions to assist with
mental health (COEH) - critical appraisal.http://research.bmh. the critical appraisal of cross-sectional studies 2018. Available at
manchester.ac.uk/epidemiology/COEH/undergraduate/ https://www.cardiff.ac.uk/__data/assets/pdf_file/0010/1142974/SURE-
specialstudymodules/criticalappraisal/Accessed 1 August 2020 CA-form-for-Cross-sectional_2018.pdf. Accessed August 1, 2020.
[25] Downs SH, Black N. The feasibility of creating a checklist for the [42] Sanderson S, Tatt ID, Higgins JP. Tools for assessing quality and suscep-
assessment of the methodological quality both of randomised and tibility to bias in observational studies in epidemiology: a systematic re-
non-randomised studies of health care interventions. J Epidemiol view and annotated bibliography. Int J Epidemiol 2007;36:666e76.
Community Health 1998;52:377e84. [43] Jarde A, Losilla JM, Vives J. Methodological quality assessment
[26] DuRant RH. Checklist for the evaluation of research articles. J Ado- tools of non-experimental studies: a systematic review. An Psicol
lesc Health 1994;15:4e8. 2012;28:617e28.