Professional Documents
Culture Documents
See the Terms and Conditions (https://onlinelibrary.wiley.com/terms-and-conditions) on Wiley Online Library for rules of use; OA articles are governed by the applicable Creative Commons License
|
Received: 26 July 2020 Accepted: 20 October 2020
DOI: 10.1111/1365-2664.14153
FORUM
1
UK Centre for Ecology & Hydrology,
Oxfordshire, UK Abstract
2
Evidence Synthesis Lab, School of Natural 1. Variable study quality is a challenge for all the empirical sciences, but perhaps
and Environmental Science, University of
Newcastle, Newcastle-upon-Tyne, UK
particularly for disciplines such as ecology where experimentation is frequently
hampered by system complexity, scale and resourcing. The resulting heteroge-
Correspondence
Oliver L. Pescott
neity, and the necessity of subsequently combining the results of different study
Email: oliver.pescott@ceh.ac.uk designs, is a fundamental issue for evidence synthesis.
KEYWORDS
causal inference, epidemiology, evidence synthesis, meta-analysis, meta-regression, multilevel
modelling, quality scoring, study design
This is an open access article under the terms of the Creative Commons Attribution License, which permits use, distribution and reproduction in any medium,
provided the original work is properly cited.
© 2022 The Authors. Journal of Applied Ecology published by John Wiley & Sons Ltd on behalf of British Ecological Society.
13652664, 2022, 5, Downloaded from https://besjournals.onlinelibrary.wiley.com/doi/10.1111/1365-2664.14153 by CAPES, Wiley Online Library on [11/03/2023]. See the Terms and Conditions (https://onlinelibrary.wiley.com/terms-and-conditions) on Wiley Online Library for rules of use; OA articles are governed by the applicable Creative Commons License
1188
Journal of Applied Ecology Pescott and Stewart
13652664, 2022, 5, Downloaded from https://besjournals.onlinelibrary.wiley.com/doi/10.1111/1365-2664.14153 by CAPES, Wiley Online Library on [11/03/2023]. See the Terms and Conditions (https://onlinelibrary.wiley.com/terms-and-conditions) on Wiley Online Library for rules of use; OA articles are governed by the applicable Creative Commons License
Pescott and Stewart 1189
Journal of Applied Ecology
developing arguments put forward by Greenland (1994a, 1994b), used in ways that are known to avoid, or minimise, the biases as-
noted that while such quality score weightings may reduce bias, sociated with the use of unidimensional quality scores that subsume
they will often do this at the expense of increased variance in the or ignore the detail of primary studies under review (Greenland &
weighted estimator. Overall inference may therefore not be im- O’Rourke, 2001; Turner et al., 2009). Even studies that appear to
proved. Greenland and O’Rourke (2001) pointed out that in this situ- be high quality may still contain non-obvious biases, and apparently
ation, methods that formally trade-off bias and variance to minimise lower quality studies could in fact be unbiased for the effect of inter-
overall error, such as hierarchical meta-regression, are likely to be est. Ultimately, corrections for study bias rely on arguments that may
more appropriate (Greenland, 2000; see also Moreno et al., 2012). not be empirically provable in any given case: context and expert
Regression-based approaches to study quality adjustment also judgement must be used to assess whether or not claims for study-
have the advantage of being able to adjust summary estimates specific bias are either corroborated by available evidence, plausi-
for the directions in which different biases might act (Greenland ble but uncorroborated, or implausible (Greenland & Pearce, 2015).
& O’Rourke, 2001), another reason why they are recommended Subsequently adjusting for study-level drivers of bias is the cur-
by modern epidemiological bias assessment frameworks (Higgins rent best practice in epidemiology (Greenland & O’Rourke, 2008;
et al., 2011; Shea et al., 2017). Thompson et al., 2011), a science that has already grappled with
Greenland and O’Rourke (2001) outlined various meta- this issue for several decades. We suggest that ecologists under-
regression approaches that could potentially overcome the limita- taking evidence synthesis focus critical appraisal on the generation
tions of unidimensional study quality scores by regressing outcomes of transparent value judgements about causation, the potential for
on individual study ‘quality item’ covariates. This approach is based confounding (risk of bias) and the appropriateness of the outcome
on the suggestion of Rubin (1990, 1992) that using a regression measure (directness), using sensitivity analyses and hierarchical
framework to extrapolate to an ‘ideal’ study is theoretically justi- meta-regression to quantify and model uncertainty arising from
fied in preference to estimating an average effect from the literature study quality, rather than automatically using generic weights. The
(Moreno et al., 2012). One could argue that the simulation-based considerable risk presented by such weighting processes to thought-
accuracy weights of CEA could be used as a quality item in a hi- ful ecological synthesis should be clear.
erarchical meta-regression framework; however, summary quality
scores resulting from generic simulation approaches are unlikely to AU T H O R S ' C O N T R I B U T I O N S
adequately represent the study-specific, multi-dimensional nature O.L.P. and G.B.S. conceived the ideas; O.L.P. led the writing of the
of bias. The accuracy weight approach of CEA necessarily combines, manuscript. Both authors contributed critically to the drafts and
and therefore ignores, many different and separate features of study gave final approval for publication.
quality. For example, as discussed above, the weights derived by
CEA are dependent on the particular effect size metrics calculated, DATA AVA I L A B I L I T Y S TAT E M E N T
on assumptions about balance across different treatment arms, and We will not be archiving data because this manuscript does not use
on whether or not the parameter values and distributions used in data.
any simulation are truly appropriate for any set of studies to which
the resulting weights might be applied. They also ignore issues relat- ORCID
ing to external validity (i.e. the representativeness of a study relative Oliver L. Pescott https://orcid.org/0000-0002-0685-8046
to its inferential target population; Turner et al., 2009) that will be
important for some types of evidence synthesis. The individual sim- REFERENCES
ulation parameters of CEA, and the weights themselves, collapse the Christie, A. P., Amano, T., Martin, P. A., Shackelford, G. E., Simmons,
multiple dimensions of individual study bias into a single score of a B. I., & Sutherland, W. J. (2019). Simple study designs in ecol-
ogy produce inaccurate estimates of biodiversity responses.
type that is highly unlikely to be perfectly proportional to the actual
Journal of Applied Ecology, 56(12), 2742–2754. https://doi.
overall biases exhibited across and within the study design types org/10.1111/1365-2664.13499
(Greenland & O’Rourke, 2001). The use of the CEA accuracy weights Deeks, J. J., Dinnes, J., D’Amico, R., Sowden, A. J., Sakarovitch, C., Song, F.,
as a covariate in a meta-regression framework would therefore likely Petticrew, M., & Altman, D. G. (2003). Evaluating non-randomised
intervention studies. Health Technology Assessment, 7(27), 1–173.
still result in bias. For this reason, Greenland and O’Rourke (2001,
Glymour, M. M., Weuve, J., Berkman, L. F., Kawachi, I., & Robins, J. M.
2008) and others (Higgins et al., 2011; Shea et al., 2017) recommend (2005). When is baseline adjustment useful in analyses of change?
keeping quality items separate and estimating them individually for An example with education and cognitive change. American Journal
each study included in a meta-analysis. of Epidemiology, 162(3), 267–278.
Greenland, S. (1994a). A critical look at some popular meta-analytic
methods. American Journal of Epidemiology, 140(3), 290–296.
https://doi.org/10.1093/oxfordjournals.aje.a117248
2 | CO N C LU S I O N Greenland, S. (1994b). Quality scores are useless and potentially mislead-
ing. American Journal of Epidemiology, 140(3), 300–3 01.
Detailed assessments of individual study validity are most likely to Greenland, S. (2000). When should epidemiologic regressions use ran-
dom coefficients? Biometrics, 56, 915–921.
provide appropriate and focused measures of quality that can be
|
13652664, 2022, 5, Downloaded from https://besjournals.onlinelibrary.wiley.com/doi/10.1111/1365-2664.14153 by CAPES, Wiley Online Library on [11/03/2023]. See the Terms and Conditions (https://onlinelibrary.wiley.com/terms-and-conditions) on Wiley Online Library for rules of use; OA articles are governed by the applicable Creative Commons License
1190
Journal of Applied Ecology Pescott and Stewart
Greenland, S., & O’Rourke, K. (2001). On the bias produced by quality Rubin, D. B. (1992). Meta-analysis: Literature synthesis or effect-size sur-
scores in meta-analysis, and a hierarchical view of proposed solu- face estimation? Journal of Educational Statistics, 17, 363–374.
tions. Biostatistics, 2(4), 463–471. https://doi.org/10.1093/biost Rubin, D. B. (2007). The design versus the analysis of observational
atistics/2.4.463 studies for causal effects: Parallels with the design of randomized
Greenland, S., & O’Rourke, K. (2008). Meta-analysis. In K. J. Rothman, trials. Statistics in Medicine, 26(1), 20–36. https://doi.org/10.1002/
S. Greenland, & T. L. Lash (Eds.), Modern epidemiology (3rd ed., pp. sim.2739
652–683). Lippincott Williams & Wilkins. Senn, S. J. (1989). Covariate imbalance and random allocation in clinical
Greenland, S., & Pearce, N. (2015). Statistical foundations for model- trials. Statistics in Medicine, 8(4), 467–475. https://doi.org/10.1002/
based adjustments. Annual Review of Public Health, 36, 89–108. sim.478008 0410
Higgins, J. P. T., Altman, D. G., Gøtzsche, P. C., Jüni, P., Moher, D., Oxman, Senn, S. J. (2005). An unreasonable prejudice against modelling?
A. D., Savović, J., Schulz, K. F., Weeks, L., Sterne, J. A. C., Cochrane Pharmaceutical Statistics, 4, 87–89.
Bias Methods Group, & Cochrane Statistical Methods Group. (2011). Senn, S. J. (2020). Being just about adjustment. Retrieved from https://
The Cochrane Collaboration’s tool for assessing risk of bias in ran- errors tatis tics.com/2020/03/16/stephe n-s enn-b eing-just-a bout
domised trials. BMJ, 343, d5928. https://doi.org/10.1136/bmj.d5928 -adjustment-guest-post/
Lortie, C. J., Stewart, G., Rothstein, H., & Lau, J. (2015). How to critically Shea, B. J., Reeves, B. C., Wells, G., Thuku, M., Hamel, C., Moran, J.,
read ecological meta-analyses. Research Synthesis Methods, 6(2), Moher, D., Tugwell, P., Welch, V., Kristjansson, E., & Henry, D. A.
124–133. https://doi.org/10.1002/jrsm.1109 (2017). AMSTAR 2: A critical appraisal tool for systematic reviews
Moreno, S. G., Sutton, A. J., Thompson, J. R., Ades, A. E., Abrams, K. R., that include randomised or non-randomised studies of healthcare
& Cooper, N. J. (2012). A generalized weighting regression-derived interventions, or both. BMJ, 358, j4008. https://doi.org/10.1136/
meta-analysis estimator robust to small-study effects and hetero- bmj.j4008
geneity. Statistics in Medicine, 31(14), 1407–1417. Thompson, S., Ekelund, U., Jebb, S., Lindroos, A. K., Mander, A., Sharp,
Pescott, O. L., & Stewart, G. B. (2014). Assessing the impact of human tram- S., Turner, R., & Wilks, D. (2011). A proposed method of bias ad-
pling on vegetation: A systematic review and meta-analysis of exper- justment for meta-analyses of published observational studies.
imental evidence. PeerJ, 2, e360. https://doi.org/10.7717/peerj.360 International Journal of Epidemiology, 40(3), 765–777. https://doi.
Roberts, C., & Torgerson, D. J. (1999). Baseline imbalance in randomised org/10.1093/ije/dyq248
controlled trials. BMJ, 319(7203), 185. https://doi.org/10.1136/ Turner, R. M., Spiegelhalter, D. J., Smith, G. C. S., & Thompson, S. G.
bmj.319.7203.185 (2009). Bias modelling in evidence synthesis. Journal of the Royal
Rubin, D. B. (1990). A new perspective. In K. W. Wachter & M. L. Straf Statistical Society: Series A (Statistics in Society), 172(1), 21–47.
(Eds.), The future of meta-analysis (pp. 155–166). Russell Sage. https://doi.org/10.1111/j.1467-985X.2008.00547.x