35
Sample size issues & Trial Quality David Torgerson

Sample size issues & Trial Quality David Torgerson

Embed Size (px)

Citation preview

Page 1: Sample size issues & Trial Quality David Torgerson

Sample size issues & Trial Quality

David Torgerson

Page 2: Sample size issues & Trial Quality David Torgerson

Chance• When we do a trial we want to be sure that any

effect we see is not simply by chance.• The probability of any difference occurring by

chance declines as the sample size increases.• Also as the sample size increases smaller

differences are likely to be statistically significant.

Page 3: Sample size issues & Trial Quality David Torgerson

Statistical significance• A p value of 0.05 means if we repeated

the same trial 100 times we would expect to observe the difference that occurred by chance 5 times.

• A small p value does not relate to the strength of the association a small difference will have a small p value if the sample size is large.

Page 4: Sample size issues & Trial Quality David Torgerson

Power• As well as significance there is the issue of

power.• A sample size might have 80% power to

detect a specified difference at 5%. In other words for a given sample size we would have an 80% probability of observing an effect if it exists with a 5% significance.

Page 5: Sample size issues & Trial Quality David Torgerson

Sample Size

• Don’t believe statistics textbooks. A trial can NEVER be too big. Most trials in education use tiny sample sizes (e.g., 30 participants).

• A small trial will miss an important difference. For example, if a school based intervention increased exam pass rates by 10% this would have very important benefits to society. BUT we would need a trial of at lease 800 participants in an individually randomised trial to observe this effect.

Page 6: Sample size issues & Trial Quality David Torgerson

Sample Size

• Small trials will miss important differences.

• Bigger is better in trials.

• Why was the number chosen? For example “given an incidence of 10% we wanted to have 80% power to show a halving to 5%” or “we enrolled 100 participants”.

Page 7: Sample size issues & Trial Quality David Torgerson

What is a reasonable difference?

• In a review by Lipsey and Wilson in 1993 of all quasi-experiments in the social sciences they found few effective interventions had effect sizes greater than 0.5. Health care produces similar gains of 0.5 of a standard deviation or lower.

Lipsey & Wilson, 1993. American Psychologist 48, 1181-1209

Page 8: Sample size issues & Trial Quality David Torgerson

Effect size & Sample size• We should, therefore, plan trials that are

large enough to identify a difference of 0.5 of a standard deviation between the two experimental groups if it exists.

Page 9: Sample size issues & Trial Quality David Torgerson

Who needs a statistician?• A simple way to calculate a sample size is

to take 32 (for 80% power, 42 for 90%) and divide this by the square of effect size.

• 0.5 squared is 0.25. 32/0.25 = 128.

• 0.25 is 512, note halving the effect size quadruples the sample size.

Page 10: Sample size issues & Trial Quality David Torgerson

Cluster sample size• Because often in education we will

randomise by classes or schools we need to take the correlation between pupils into account for sample size calculations. This can often lead to a doubling or more of the sample participants.

Page 11: Sample size issues & Trial Quality David Torgerson

Reporting Quality of Trials

• As well as having an adequate sample size there are other important aspects of trial quality.

Page 12: Sample size issues & Trial Quality David Torgerson

Important quality items

• Allocation method;» method randomisation;» secure randomisation.

• Intention to treat analysis.

• Blinding.

• Attrition.

Page 13: Sample size issues & Trial Quality David Torgerson

Blinding

• Who knew who got what when?

• Was the participant blind?

• Most IMPORTANT was outcome assessment blind?

Page 14: Sample size issues & Trial Quality David Torgerson

Attrition

• What was the final number of participants compared with the number randomised?

• What happened to those lost along the way?

• Was there equal attrition?

Page 15: Sample size issues & Trial Quality David Torgerson

External Validity

• Once a trial has high internal validity our next task is to assess whether its results are applicable outside its sample.

• Are participants similar to the general population on who we would apply the intervention?

• Was intervention used generalisable?

Page 16: Sample size issues & Trial Quality David Torgerson

Methods comparison of trials

• We undertook a methodological review of RCTs in health and education to answer the following questions:» Were only bad trials prevalent in health care?» Was methodological quality improving over

time?

Torgerson et al. BERJ, 2004: Accepted.

Page 17: Sample size issues & Trial Quality David Torgerson

Study Characteristics

Characteristic Drug Health Education Cluster Randomised 1% 36% 18% Sample size justified 59% 28% 0% Concealed randomisation 40% 8% 0% Blinded Follow-up 53% 30% 14% Use of CIs 68% 41% 1% Low Statistical Power 45% 41% 85%

Page 18: Sample size issues & Trial Quality David Torgerson

Change in concealed allocation

05

101520253035404550

Drug No Drug

<1997>1996

NB No education trial used concealed allocation

P = 0.04 P = 0.70

Page 19: Sample size issues & Trial Quality David Torgerson

Blinded Follow-up

0

10

20

30

40

50

60

Drug Health Education

<1997>1996

P = 0.03P = 0.13P = 0.54

Page 20: Sample size issues & Trial Quality David Torgerson

Underpowered

0

10

20

30

40

50

60

70

80

90

Drug Health Education

<1997>1996

P = 0.22P = 0.76P = 0.01

Page 21: Sample size issues & Trial Quality David Torgerson

Mean Change in Items

0

0.5

1

1.5

2

2.5

3

3.5

Drug No Drug Education

<1997>1996

P= 0.001 P= 0.07 P= 0.03

Page 22: Sample size issues & Trial Quality David Torgerson

Quality conclusions

• Apart from ‘drug’ trials the quality of health care trials is poor and not improving outside of major journals.

• Education trials are bad and getting worse!

Page 23: Sample size issues & Trial Quality David Torgerson

Trial Examples

• CBT treatment vs Fire Safety Education (FSE) for child arsonists.

• N = 38 boys randomised to receive FSE or CBT.

• Outcomes measured at 13 weeks and 12 months included firesetting and match play behaviour

Kolko J Child Psychiatr 2001:42:359.

Page 24: Sample size issues & Trial Quality David Torgerson

Results

• Outcomes were mixed in some outcomes were favourable for CBT, whilst for others, there was no difference.

Page 25: Sample size issues & Trial Quality David Torgerson

Results

0102030405060708090

100

FSE CBT

Page 26: Sample size issues & Trial Quality David Torgerson

Problems with Trial

• Too SMALL.• Trial could have missed very important

differences.• Outcomes were NOT arson, there were no

reports of arson by any children in the study.

• Unsure whether randomisation was concealed.

Page 27: Sample size issues & Trial Quality David Torgerson

Domestic violence experiment

• 404 men convicted of partner abuse were randomised to probation or counselling.

• Data were collected at 12 months on re-arrests, beliefs and behaviours on partner abuse.

Feder & Dugan 2002; Justice Quarterly 19;343.

Page 28: Sample size issues & Trial Quality David Torgerson

Results

• No difference in re-offending as measured by re-arrest statistics (I.,e 24% in both groups).

• No differences in attitudes towards partner abuse.

Page 29: Sample size issues & Trial Quality David Torgerson

Trial Methods

• Trial was relatively large (> 200 in each group) would have had enough power to detect a halving of offending. That is 24% down to 12%.

• For ‘beliefs’ there was a high drop-out rate, (50%) which may make those results unreliable.

• Allocation appeared to be secure.• Cross-over was slight.• Unclear as to whether re-arrest data was collected

‘blindly’.

Page 30: Sample size issues & Trial Quality David Torgerson

Conclusion

• Counselling is probably an ineffective method of trying to prevent spousal abuse.

• Other interventions should be sought.

• Message: if you’re being battered by your Spouse, don’t bother with counselling!

Page 31: Sample size issues & Trial Quality David Torgerson

Preventing unscheduled school transfers

• Unscheduled school transfers are associated with poor academic outcomes. The raising healthy children project in Seattle aimed to put into place interventions among high risk students who exhibit academic or behavioural problems.

Fleming et al, 2001: Evaluation Review 25:655.

Page 32: Sample size issues & Trial Quality David Torgerson

Design• Cluster randomised trial

• 10 schools randomised. 5 experimental schools received a variety of interventions to help high risk students and their families.

• Analysis was multilevel to take into account clustering.

Page 33: Sample size issues & Trial Quality David Torgerson

Results

• The intervention showed that there was a reduction of 2/3rds in the transfer rates, which was statistically significant. 61% versus 45% difference 16% NOTE, that the intervention schools still had a high transfer rate.

• Also effects of intervention waned over the 5 years, suggesting it would need to be continuous to be effective.

Page 34: Sample size issues & Trial Quality David Torgerson

Study implications

• Study showed an effective intervention. Number of clusters (10) was on small side, ideally should have been more. High chance of missing a smaller effect.

Page 35: Sample size issues & Trial Quality David Torgerson

Conclusions

• RCT is the BEST evaluative method.

• They can, and have been done, in the field of education.

• We need MORE larger and better quality trials to inform future policy in this area.