Professional Documents
Culture Documents
READ THIS SERIES IF YOU WANT TO yy BE MORE EVIDENCEBASED IN YOUR PRACTICE yy FEEL MOTIVATED TO READ JOURNAL ARTICLES yy INFLUENCE DEVELOPMENT OF YOUR SERVICE
he intervention study tool we will explore here is suitable for appraising experimental studies which set out to measure outcomes from provision of a therapy (or other kind of intervention) whose purpose is to alleviate either a persons difficulties or their impact on the persons life. We have now covered several appraisal tools, so you may be feeling it is not always obvious which to choose. To help you sort out what a study is really about, and whether it refers to interventions, try to match its aim to one of these questions: 1. What can we observe in people who have a particular risk factor or condition? If the investigators did not provide an intervention themselves but were looking at outcomes in people who may have received one type of intervention or another, you are probably looking at an observational design. In observational studies, the researchers do not provide an intervention or otherwise seek to promote change in individuals. They examine aspects of peoples past and/or present life and seek to find factors that are associated with each other. Sometimes this may involve collecting data about the intervention(s) received. The underlying quest in observational studies is usually to identify factors which might be causally related (and thus may be incorporated into interventions to produce better outcomes in the future). We will come back to the appraisal of observational studies in a future Journal Club. 2. What can people tell us about their experience of living with a particular risk factor or condition? Does the study appear to be less about objective observation of outcomes and facts, and more about eliciting peoples subjective experiences of the impact of their condition or of being the recipient of interventions? You may find that the qualitative study appraisal tool in Journal Club 2 (Winter 2010) is most applicable. 3. Does a given therapy or intervention work? If a quick skim of the abstract and/or method of the article suggests that the author is simply describing an intervention (or range of interventions), it does not qualify as an intervention study. Instead, it may be appraised using the expert opinion tool in Journal Club 1 (Autumn 2010). If the authors are reviewing or combining the results of more than one published study, then the systematic review
tool in Journal Club 3 (Spring 2011) is more likely to be the one to use. Once you have ascertained that the study in question is indeed an intervention study, you then need to ask one further question to drill down to the right tool: 4. Does this study use a group design or a single-subject design? Did the researchers provide the intervention to one group of people and compare the group results with those of people who either received a different intervention or no intervention at all? If the answer is yes, you can use the intervention study tool in this article. If, on the other hand, the study is about one persons treatment results or a series of individuals in which there is no comparison group receiving a different or no treatment, then you are probably dealing with a single-subject design. An appraisal tool for single-subject design studies will follow in a subsequent Journal Club. Single-subject designs most often use multiple baselines to compare treated and untreated targets within a single individual, so look out for that terminology. The gold standard design for an intervention study is the randomised controlled trial (RCT). The framework in this article can also be used for other types of intervention study, such as exploratory, preliminary or pilot-study designs, but you need to use the separate tool for appraising reports of single-case study and case series designs. This is because group designs help minimise the opportunity for bias arising from the Hawthorne effect (people tend to behave or respond differently when they know they are participants in a study) and the placebo effect (people receiving treatment of any kind do better than those receiving no treatment). It is not possible to rule out these effects from research using single-subject designs. Group studies in which the data from individuals is pooled but there is no control (comparison) group can pretty much be avoided as inadequate. Since neither Hawthorne nor placebo effect are controlled, any treatment outcomes could be entirely non-specific. There is little point in demonstrating a change that could have resulted from any old intervention Pam
Enderbys description of early aphasia therapy research as trials of random niceness comes to mind again. One exception may be when research is at the pre-clinical or modelling phase of the development of a complex intervention, when you may be interested in the content of and theoretical justification for untested interventions (MRC, 2000). Having identified your intervention study, ask the following questions to appraise its quality. The issues are the ones that you will find in any of the published frameworks for appraisal of randomised controlled trials, such as Critical Appraisal Skills Programme (CASP) (PHRU, 2006) or the PEDro Scale (Physiotherapy Evidence Database, 1999). And if you want a good laugh (yes, really!) about why these issues are important, I recommend Ben Goldacres Bad Science (2009). Beware, though, as you may never feel the same way again about vitamin supplements, brain gym or homeopathy This appraisal starts the same way as for a systematic review (Journal Club 3, Spring 2011): Question 1: What question was being asked, and is this an important clinical question?
Try formulating the reviewers stated aims into a research question if they have not done so explicitly in the article. Is the question clearly focused in terms of the PICO framework: Population studied Intervention given Comparator/control(s) used Outcomes considered? Is this question important for your clinical practice? If the reviewers question does not quite fit the bill, what question(s) do you wish they had asked instead? (The study may provide a partial answer.)
18
Do the authors discuss why this study was carried out, either as a randomised controlled trial or otherwise? Most of the interventions of interest to speech and language therapists are complex interventions. Medical Research Council guidelines define these as ones which, whether therapeutic or preventative, comprise a number of separate elements which seem essential to the proper functioning of the intervention although the active ingredient of the intervention that is effective is difficult to specify. If we were to consider a randomised controlled trial of a drug vs. a placebo as being at the simplest end of the spectrum, then we might see a comparison of a stroke unit to traditional care as being at the most complex end of the spectrum. The greater the difficulty in defining precisely what, exactly, are the active ingredients of an intervention and how they relate to each other, the greater the likelihood that you are dealing with a complex intervention. (MRC, 2000, p.1) Just because the randomised controlled trial design is the gold standard for interventions does not mean researchers can or should be undertaking a randomised controlled trial of a complex intervention before the preliminary work has been done to establish an adequate description of its components: Complex interventions are built up from a number of components, which may act both independently and inter-dependently. The components usually include behaviours, parameters of behaviours (e.g. frequency, timing), and methods of organising and delivering those behaviours (e.g. type(s) of practitioner, setting and location). It is not easy precisely to define the active ingredients of a complex intervention. For example, although research suggests that stroke units work, what, exactly, is a stroke unit? What are the active ingredients that make it work? The physical set-up? The mix of care providers? The skills of the providers? The technologies available? The organisational arrangements? (MRC, 2000, p.2) So you need to ask at what stage of development this intervention is and, therefore, whether a randomised controlled trial at this stage is appropriate. In one of our local journal clubs, we reviewed an article on Constraint Induced Language Therapy (CILT) for people with aphasia (Maher et al., 2006) (figure 1). We thought their question was something like, Does Constraint Induced Language Therapy change the spoken
Figure 2 Enough jargon to sink that little pilot boat... Prospective means looking forward so its trying to answer the question, What will happen if we do x? rather than, What happened to these people in the past? (retrospective). Its use here seems a bit redundant since this is an intervention rather than an observational study. Repeated measures means you use the same tools to measure participants skills, behaviours etc. at pre-set intervals of time before, sometimes during, and after the treatment. This is one of the methods usually associated with a single-case study or case series design. This study used a mixture of standardised measures such as the Western Aphasia Battery and Boston Naming Test, and narrative analysis of a retelling of the Cinderella story, but no measures to detect more functional gains. Pilot study is using the analogy of the pilot boat which leads the way for a larger, less manoeuvrable (but more powerful) ship. Describing it as a pilot study also releases the authors from some of the constraints of the randomised controlled trial, allowing them to get a study into print that would otherwise fall foul of peer reviewers criticisms.
language behaviour of people with chronic, stroke-induced aphasia (with or without dyspraxia) compared to a similar dosage of multi-modal PACE (Promoting Aphasics Communicative Effectiveness)-type therapy? Maybe not the question we would have chosen to ask but still potentially of interest. The design was described by the authors as a prospective, repeated measures pilot study (figure 2). Not a randomised controlled trial then, but probably an appropriate design for early studies of a new therapy. Question 3: Were participants appropriately allocated to intervention and control groups?
Was the method of allocation adequately described and truly random? Each participant should have had an equal chance of being allocated to any of the study groups. It might seem a bit pedantic, putting in all that detail of sealed envelopes and independent personnel, but there is a reason for it. The whole purpose of a group study is to test whether the intervention effect you are interested in occurs when the intervention is provided for a particular
population of people with communication or swallowing difficulties. Randomisation ensures that groups are sufficiently similar to wipe out any differences in results that might occur because of individual differences. Have you ever read the results of a group study and been frustrated at the lack of detail about its impact on individuals? Well, thats exactly the point. The group study should tell you about the strength of the interventions effect irrespective of individual differences, so the study method needs to control for possible interference from inadequate randomisation. There is a large accumulation of evidence that people are influenced by knowledge of study participants and so (even if only subconsciously) may allocate to treatment rather than control groups participants who are, for example, keener, more worthy, more likely to adhere to the treatment schedule Hence the need for distance, independence and no personal contact with the participants. Researchers may use stratification to ensure equal or balanced numbers of certain types of participant such as boys and girls, or people living in three different parts of the country, or people with different levels of severity of a condition. Randomisation then takes place within each stratification group. Were the groups well balanced? Look for tables or other data describing potentially important characteristics of the study groups, and evidence that the authors have checked (with statistical tests if appropriate) that the groups are indeed similar enough to control for
19
As in other appraisal frameworks you need to examine the detail to determine whether there are important differences between the population sampled for the research study and the population of your own context. It may also be useful to consider practical aspects, such as could you provide the same intervention in your setting? Evidence-based practice is often modelled as a triangle with its three components being best research evidence + expertise of the practitioner + client preference. As the impact on clinical decision-making of resources and service delivery options arising from the local context cannot be ignored, some authors include these constraints as a fourth component of evidence-based practice. Question 6: Was the selection and follow-up of the participants adequate?
Blinding is about minimising the potential for observer bias, specifically the tendency for assessors to expect more change amongst individuals in treatment than control groups. The existence of this effect, like the placebo, is very well substantiated in the medical and psychological literature, so the need for blinding is an important aspect of good research design. However, there are limitations on where and how much speech and language therapy treatment studies can employ blinding. In a single blind study, the participants do not know which group they are in, while in a double blind study, neither do the investigators. In speech and language therapy research, it is very unlikely that participants and practitioners can be blind to participants group, so it is good practice for other people to undertake the assessments and outcome measures. A bit of common sense can help. Did they employ blinding wherever it was possible to do so? If blinding was missing from some measures, does it really matter for this study? No information was given in Maher et al. (2006) on who administered the standardised measures, but the narrative sample analysis was done blind to both treatment group and whether pre-, post- or 1 month after treatment status. It seemed to us much more important that blinding was applied to these measures, so we were less concerned about the potential for observer bias from possible lack of blinding in the standardised tests.
Were there any differences in data collection that might introduce performance bias? For example, were there any differences between the groups in the time intervals between measures? Did assessments undertaken by participants differ so that one group received a lot more attention from researchers, therapists or other practitioners? Did the study have enough participants to minimise the play of chance? Look for a report of a power calculation if this is a fullscale randomised controlled trial. Power calculations estimate how many participants are needed to be reasonably sure of finding out whether an intervention really does have a specific effect. If you cant handle the stats, at least check that the authors have considered the possibility of having too few participants. The danger of this is that the
Could you replicate the intervention on the basis of the information given? Look back to the earlier quote from the Medical Research Council Guidelines. The authors should have provided a thorough description of the intervention with reference to components such as: intensity (eg. frequency and length of sessions) dosage (eg. how many hours of therapy over what length of time) scheduling (eg. distributed learning or massed practice) content (eg. targets, success criteria, resources) role/responses of therapist or other practitioner (how changes in behaviour are facilitated eg. modelling, contingent responses, scaffolding, recasting) methods used to minimise systematic differences between practitioners delivering the interventions (eg. manualising the intervention, training the practitioners, measuring fidelity). The authors should provide information on the track record of the intervention, if it has one, with a summary of the current evidence for its efficacy (delivery under ideal conditions) or effectiveness (delivery in real world settings). For innovative interventions, ask whether they are founded on sound theoretical and practice-based principles.
Critical appraisal for speech and language therapists (CASLT) Download the intervention study framework document from www.speechmag.com/Members/CASLT for use by yourself or with colleagues in a journal club.
20
SPEECH & LANGUAGE THERAPY IN PRACTICE SUMMER 2011
Could the people included in the study be different from your population in ways that would produce different results? Does your local setting differ much from that of the trial? Could you provide the same treatment in your setting, given local constraints and opportunities? If not, what changes would be needed to your services resources or to their deployment? Consider outcomes from different points of view including that of the client and family, the wider community, professionals and policy makers. These were our comments on Maher et al. (2006): We remain unconvinced about how replicable this conceptualisation of Constraint Induced Language Therapy would be in our settings getting well-matched dyads or triads would be difficult in most settings. How would you do this, for example, with someone in a care home? We thought there might be potential for delivering massed practice intensity via telemedicine or computer, although this would need a different core of therapy tasks. We worried how you would retrieve a negative outcome from using Constraint Induced Language Therapy with a client in order to get back to a total communication approach. We
21