50
How much confidence can we have in published medical research? Doug Altman Centre for Statistics in Medicine University of Oxford

Doug Altman 15 Jan09 V4

Embed Size (px)

DESCRIPTION

How much confidence do we have in published...

Citation preview

Page 1: Doug Altman 15 Jan09 V4

How much confidence can we have in published medical

research?

Doug Altman

Centre for Statistics in MedicineUniversity of Oxford

Page 2: Doug Altman 15 Jan09 V4

Baltimore, January 2009 2

Richard HortonEditor, Lancet

“Can you trust the research published in medical journals? Mostly I believe you can. But not always … some of the most newsworthy reports published by our leading journals spontaneously combust under the gentlest of scrutiny.”

Guardian 16 Sept 2004

Page 3: Doug Altman 15 Jan09 V4

Baltimore, January 2009 3

John Ioannidis(Epidemiologist, Greece)

“Most research findings are false for most research designs and for most fields”

[PLoS Medicine 2005]

Page 4: Doug Altman 15 Jan09 V4

Baltimore, January 2009 4

Research and publication

Medical research should advance scientific knowledge and – directly or indirectly – lead to improvements in treatment or prevention of disease

A research report is usually the only tangible evidence that the study was done, how it was done, and what the findings were

Scientific manuscripts should present sufficient data so that the reader can fully evaluate the information and reach his or her own conclusions about results– Assess reliability and relevance

Page 5: Doug Altman 15 Jan09 V4

Baltimore, January 2009 5

The purpose of a research article

Clinicians might read it to learn how to treat their patients better– “Editors, reviewers and authors are often tempted to

pander to this group, by sexing up the results with unjustified clinical messages – sometimes augmented by an even more unbalanced press release.” [Buckley Emerg Med Australas 2005]

Researchers might read it to help plan a similar study or as part of a systematic review– Need a clear understanding of exactly what was done

Page 6: Doug Altman 15 Jan09 V4

Baltimore, January 2009 6

Some major causes of bias

Incorrect methods

Poor reporting– Failure to disclose key aspects of how the study was

done – Selective reporting of only some of a large number of

results• Groups, outcomes, subgroups, etc

Errors of interpretation

Page 7: Doug Altman 15 Jan09 V4

Baltimore, January 2009 7

3 underlying major problems in medical research

Unclear main study question

Multiplicity

Over-reliance on significance tests and the holy grail of P<0.05

… very often in combination

Page 8: Doug Altman 15 Jan09 V4

Baltimore, January 2009 8

Composite outcomes

Lotrafiban vs placebo in coronary and cerebrovascular disease [Topol et al, Circulation 2003]

Page 9: Doug Altman 15 Jan09 V4

Baltimore, January 2009 9

Post hoc specification

Choosing the analysis after looking at the data– A specific form of incomplete reporting – Many other analyses are implied but not carried out

Data-derived analyses are misleading– can have a huge effect

Page 10: Doug Altman 15 Jan09 V4

Baltimore, January 2009 10

Vasopressin vs epinephrine for out-of-hospital cardiopulmonary resuscitation[Wenzel et al, NEJM 2004]

Primary outcome – hospital admission (alive) The overall comparison gives

OR = 1.26 (95% CI 0.98 to 1.62) [P=0.06 ]

Page 11: Doug Altman 15 Jan09 V4

Baltimore, January 2009 11

Wenzel et al, NEJM 2004

Page 12: Doug Altman 15 Jan09 V4

Baltimore, January 2009 12

Wenzel et al, NEJM 2004

No mention in Methods of planning to look at subgroups

Comparison of P values is wrong No significant difference between 3 groups

(interaction test)

Page 13: Doug Altman 15 Jan09 V4

Baltimore, January 2009 13

Comparing trial publications with protocols

Comparison of protocols and publications– 102 RCTs submitted to a Danish Ethics committee in

1994-95 • Protocols and subsequent journal articles (122 articles)• Questionnaire sent to all authors

Frequent discrepancies between the protocol and the published report– Specification of primary outcomes: 51/82 (62%)

[Chan et al, JAMA 2004]

Page 14: Doug Altman 15 Jan09 V4

Baltimore, January 2009 14

[Chan et al, JAMA 2004]

Page 15: Doug Altman 15 Jan09 V4

Baltimore, January 2009 15

Primary outcome:% with Score<3 at 1 year

Primary outcome:% dead/dependent at 1

year

Example

Neurology trialSurgical intervention

P0.05

Primary outcome:% with Score<3 at 1 yr

Protocol

P<0.05

Publication

Page 16: Doug Altman 15 Jan09 V4

Baltimore, January 2009 16

Comparing trial publications with protocols

Unacknowledged discrepancies between protocols and publications– sample size calculations (18/34 trials), – methods of handling protocol deviations (19/43) – missing data (39/49), – primary outcome analyses (25/42)– subgroup analyses (25/25)– adjusted analyses (23/28)

Interim analyses were described in 13 protocols but mentioned in only five corresponding publications

[Chan et al, BMJ 2008]

Page 17: Doug Altman 15 Jan09 V4

Baltimore, January 2009 17

Biased interpretation

Authors often interpret nonsignificant results as evidence of no effect, regardless of sample size

“Conclusions in trials funded by for-profit organizations may be more positive due to biased interpretation of trial results” [Als-Nielsen et al, JAMA 2003]

Page 18: Doug Altman 15 Jan09 V4

Baltimore, January 2009 18

Page 19: Doug Altman 15 Jan09 V4

Baltimore, January 2009 19Doyle et al, Arch Surg

2008

Page 20: Doug Altman 15 Jan09 V4

Baltimore, January 2009 20

“Patients who undergo liver transplantation at age 60 or above

have 1-year and 5-year survival rates similar to those of younger

patients …”

No mention that the study was of hepatitis

C positive patients

Page 21: Doug Altman 15 Jan09 V4

Baltimore, January 2009 21

Doyle et alEffect of age of donor

Age>60 was not statistically significant in the multivariate analysis

HR=3.03 (95%CI: 0.70-12.20) P = 0.12

“There are at least 25 other studies that have demonstrated that older donors are associated with worse outcomes in recipients with hepatitis C…”

Page 22: Doug Altman 15 Jan09 V4

Baltimore, January 2009 22

Importance of transparent reporting of research

Scientific manuscripts should present sufficient data so that the reader can fully evaluate the information and reach his or her own conclusions about results– Reliability and relevance

There is much evidence that this does not happen

Assessment of reliability of published articles is seriously impeded by inadequate reporting

Authors (and journals) have an obligation to ensure that research is reported adequately – i.e. transparently and completely

Page 23: Doug Altman 15 Jan09 V4

Baltimore, January 2009 23

Evidence of poor reporting

Poor reporting: key information is missing or ambiguous

There is considerable evidence that many published articles do not contain the necessary information – We cannot tell exactly how the research was done

Page 24: Doug Altman 15 Jan09 V4

Baltimore, January 2009 24

519 Randomised trials published

in December 2000

Failure to report key aspects of trial conduct:

73% Sample size calculation 55% Defined primary outcome(s)60% Whether blinded79% Method of random sequence generation82% Method of allocation concealment

[Chan & Altman Lancet 2005]

Page 25: Doug Altman 15 Jan09 V4

Baltimore, January 2009 25

Poor reporting is a serious problem for systematic reviews and clinical guidelines

“The biggest problem was the quality of reporting, which did not allow us to judge the important methodological items ...”

“Data reporting was poor. 15 trials met the inclusion criteria for this review but only 4 could be included as data were impossible to use in the other 11.”

(Reviews on Cochrane Library, accessed on 18 Sept 07)

Page 26: Doug Altman 15 Jan09 V4

Baltimore, January 2009 26

Other study types

There is most evidence for randomised trials but similar concerns apply to all types of research

Systematic reviews Phase II trials Diagnostic accuracy Epidemiology Genetic epidemiology Prognostic markers

etc

Page 27: Doug Altman 15 Jan09 V4

Baltimore, January 2009 27

Case-control studies

“The reporting of methods in the 408 identified papers was generally poor, with basic information about recruitment of participants often absent …”

“Poor reporting of recruitment strategies threatens the validity of reported results and reduces the generalisability of studies.”

[Lee et al. Br J Psychiatry 2007]

Page 28: Doug Altman 15 Jan09 V4

Baltimore, January 2009 28

Selective reporting

In addition, there is accumulating evidence of two major threats to the medical literature

Study publication bias – studies with less interesting findings are less likely to be published

Outcome reporting bias – results included within published reports are selected to favour those with statistically significant results

Page 29: Doug Altman 15 Jan09 V4

Baltimore, January 2009 29

Impact of poor reporting

Cumulative published evidence is misleading– Biased results– Methodological weaknesses may not be apparent

Adverse effects on– Other researchers – Clinicians– Patients

Page 30: Doug Altman 15 Jan09 V4

Baltimore, January 2009 30

Whose fault is poor reporting?

Poor reporting indicates a collective failure of authors, peer reviewers, and editors … on a massive scale

Researchers (authors) may not know what information to include in a report of research

Editors may not know what information should be included

What help can be given to authors?

Page 31: Doug Altman 15 Jan09 V4

Baltimore, January 2009 31

What can be done to improve research reports?

Research Publication Knowledge dissemination

Research Scientific writing guidanceconduct Journals’ Instructions to Authorsguidance

Page 32: Doug Altman 15 Jan09 V4

Baltimore, January 2009 32

What can be done to improve research reports?

Closing the gap

Research Publication Knowledge dissemination

Research Scientific writing guidanceconduct Journals’ Instructions to Authorsguidance

Reporting guidelines

Page 33: Doug Altman 15 Jan09 V4

Baltimore, January 2009 33

Transparency and reproducibility

All key aspects of how the study was done– Allow repetition (in principle) if desired

“Describe statistical methods with enough detail to enable a knowledgeable reader with access to the original data to verify the reported results.”

[International Committee of Medical Journal Editors]

– Same principle should extend to all study aspects– Only 49% of 80 consecutive reports accepted for

publication in Evidence-Based Medicine (2005-06) gave sufficient details of the treatment studied to allow clinicians to reproduce it [Glasziou et al 2007]

Reporting of results should not be misleading

Page 34: Doug Altman 15 Jan09 V4

Baltimore, January 2009 34

An early call for guidance on reporting clinical trials (1950)

“This leads one to consider if it is possible, in planning a trial, in reporting the results, or in assessing the published reports of trials, to apply criteria which must be satisfied if the analysis is to be entirely acceptable….

“A basic principle can be set up that … it is at least as important to describe the techniques employed and the conditions in which the experiment was conducted, as to give the detailed statistical analysis of results.”

[Daniels M. Scientific appraisement of new drugs in tuberculosis. Am Rev Tuberc 1950;61:751-6.]

Page 35: Doug Altman 15 Jan09 V4

Baltimore, January 2009 35

The CONSORT Statement for reporting RCTs[Moher et al, JAMA/Annals/Lancet 2001]

Minimum set of essential items necessary to evaluate the study

22 items that should be reported in a paper – Based on empirical evidence where possible

Also a flow diagram describing patient progress through the trial

Long explanatory paper (E&E)

www.consort-statement.org

Page 36: Doug Altman 15 Jan09 V4

Baltimore, January 2009 36

Goals of CONSORT

Main objective To facilitate critical appraisal and

interpretation of RCTs by providing guidance to authors about how to improve the reporting of their trials

Secondary objective To encourage and provide incentives for

researchers to conduct high-quality, unbiased randomized trials

Page 37: Doug Altman 15 Jan09 V4

Baltimore, January 2009 37

Impact of CONSORT

All leading general medical journals and hundreds of specialist journals support CONSORT – Not necessarily enforced

Adoption of CONSORT by journals is associated with improved reporting [Plint et al, Med J Aust 2006]

Official extensions– Cluster trials, non-inferiority and equivalence trials, harms,

non-pharmacological treatments, pragmatic trials, abstracts

Unofficial extensions– e.g. Acupuncture (STRICTA), Nonrandomised public health

interventions (TREND)

Page 38: Doug Altman 15 Jan09 V4

Baltimore, January 2009 38

Other guidelines

Other study types – CONSORT as a model– QUOROM (meta-analyses of RCTs) ( PRISMA)– STARD (diagnostic studies)– STROBE (observational studies)– REMARK (tumour marker prognostic studies)– …

Such guidelines are still not widely supported by medical journals or adhered to by researchers– Their potential impact is blunted

Page 39: Doug Altman 15 Jan09 V4

Baltimore, January 2009 39

STROBE

Page 40: Doug Altman 15 Jan09 V4

Baltimore, January 2009 40

Key aspects of reporting guidelines

Guidance not requirements– Journals may enforce adherence

For authors, editors, and readers

Not methodological quality“Accurate and transparent reporting is like turning the light on before you clean up a room: It doesn’t clean it for you but does tell you where the problems are.”

[Frank Davidoff, Ann Intern Med 2000]

Adherence does not guarantee a high quality study!

Page 41: Doug Altman 15 Jan09 V4

Baltimore, January 2009 41Perrone et al, Eur J Cancer 2008

Page 42: Doug Altman 15 Jan09 V4

Baltimore, January 2009 42Perrone et al, Eur J Cancer 2008

Page 43: Doug Altman 15 Jan09 V4

Baltimore, January 2009 43

EQUATOR: Enhancing the QUAlity and Transparency Of health Research

EQUATOR grew out of the work of CONSORT and other guidelines groups

Guidelines are available but not widely used Recognised the need to actively promote

guidelines EQUATOR Network

– Editors of general and specialty journals, researchers, guideline developers, medical writers

The goal: Better reporting, better reviewing, better editing

Page 44: Doug Altman 15 Jan09 V4

Baltimore, January 2009 44

www.equator-network.org

Page 45: Doug Altman 15 Jan09 V4

Baltimore, January 2009 45

EQUATOR Core Programme: Objectives

1. Provide resources enabling the improvement of health research reporting Website Courses

2. Monitor progress in the improvement of health research reporting

Achieving funding for such work is a major headache!

Page 46: Doug Altman 15 Jan09 V4

Baltimore, January 2009 46

What does the poor quality of published studies tell us about peer review?

Peer review is difficult and only partly successful

Reviewers (and editorial staff) are unable to eliminate errors in methodology and interpretation

Readers should not assume that papers published in peer reviewed journals are scientifically sound – But, many do!

Page 47: Doug Altman 15 Jan09 V4

Baltimore, January 2009 47

Some (partial) solutions

Monitoring/regulation– Ethics committees, data monitoring committees,

funders Trial registration Publication of protocols Availability of raw data Journal restrictions

Page 48: Doug Altman 15 Jan09 V4

Baltimore, January 2009 48

Reproducible research

Page 49: Doug Altman 15 Jan09 V4

Baltimore, January 2009 49

How much confidence can we have in published medical research?

Can’t rely on authors– Poor methodological quality– Poor reporting

Can’t rely on peer review– Fallible

In general, must rely on one’s own training, experience and judgement– Read with caution!– Be aware of what may be missing

Page 50: Doug Altman 15 Jan09 V4

Baltimore, January 2009 50