Upload
us-cochrane-center
View
883
Download
2
Embed Size (px)
DESCRIPTION
How much confidence do we have in published...
Citation preview
How much confidence can we have in published medical
research?
Doug Altman
Centre for Statistics in MedicineUniversity of Oxford
Baltimore, January 2009 2
Richard HortonEditor, Lancet
“Can you trust the research published in medical journals? Mostly I believe you can. But not always … some of the most newsworthy reports published by our leading journals spontaneously combust under the gentlest of scrutiny.”
Guardian 16 Sept 2004
Baltimore, January 2009 3
John Ioannidis(Epidemiologist, Greece)
“Most research findings are false for most research designs and for most fields”
[PLoS Medicine 2005]
Baltimore, January 2009 4
Research and publication
Medical research should advance scientific knowledge and – directly or indirectly – lead to improvements in treatment or prevention of disease
A research report is usually the only tangible evidence that the study was done, how it was done, and what the findings were
Scientific manuscripts should present sufficient data so that the reader can fully evaluate the information and reach his or her own conclusions about results– Assess reliability and relevance
Baltimore, January 2009 5
The purpose of a research article
Clinicians might read it to learn how to treat their patients better– “Editors, reviewers and authors are often tempted to
pander to this group, by sexing up the results with unjustified clinical messages – sometimes augmented by an even more unbalanced press release.” [Buckley Emerg Med Australas 2005]
Researchers might read it to help plan a similar study or as part of a systematic review– Need a clear understanding of exactly what was done
Baltimore, January 2009 6
Some major causes of bias
Incorrect methods
Poor reporting– Failure to disclose key aspects of how the study was
done – Selective reporting of only some of a large number of
results• Groups, outcomes, subgroups, etc
Errors of interpretation
Baltimore, January 2009 7
3 underlying major problems in medical research
Unclear main study question
Multiplicity
Over-reliance on significance tests and the holy grail of P<0.05
… very often in combination
Baltimore, January 2009 8
Composite outcomes
Lotrafiban vs placebo in coronary and cerebrovascular disease [Topol et al, Circulation 2003]
Baltimore, January 2009 9
Post hoc specification
Choosing the analysis after looking at the data– A specific form of incomplete reporting – Many other analyses are implied but not carried out
Data-derived analyses are misleading– can have a huge effect
Baltimore, January 2009 10
Vasopressin vs epinephrine for out-of-hospital cardiopulmonary resuscitation[Wenzel et al, NEJM 2004]
Primary outcome – hospital admission (alive) The overall comparison gives
OR = 1.26 (95% CI 0.98 to 1.62) [P=0.06 ]
Baltimore, January 2009 11
Wenzel et al, NEJM 2004
Baltimore, January 2009 12
Wenzel et al, NEJM 2004
No mention in Methods of planning to look at subgroups
Comparison of P values is wrong No significant difference between 3 groups
(interaction test)
Baltimore, January 2009 13
Comparing trial publications with protocols
Comparison of protocols and publications– 102 RCTs submitted to a Danish Ethics committee in
1994-95 • Protocols and subsequent journal articles (122 articles)• Questionnaire sent to all authors
Frequent discrepancies between the protocol and the published report– Specification of primary outcomes: 51/82 (62%)
[Chan et al, JAMA 2004]
Baltimore, January 2009 14
[Chan et al, JAMA 2004]
Baltimore, January 2009 15
Primary outcome:% with Score<3 at 1 year
Primary outcome:% dead/dependent at 1
year
Example
Neurology trialSurgical intervention
P0.05
Primary outcome:% with Score<3 at 1 yr
Protocol
P<0.05
Publication
Baltimore, January 2009 16
Comparing trial publications with protocols
Unacknowledged discrepancies between protocols and publications– sample size calculations (18/34 trials), – methods of handling protocol deviations (19/43) – missing data (39/49), – primary outcome analyses (25/42)– subgroup analyses (25/25)– adjusted analyses (23/28)
Interim analyses were described in 13 protocols but mentioned in only five corresponding publications
[Chan et al, BMJ 2008]
Baltimore, January 2009 17
Biased interpretation
Authors often interpret nonsignificant results as evidence of no effect, regardless of sample size
“Conclusions in trials funded by for-profit organizations may be more positive due to biased interpretation of trial results” [Als-Nielsen et al, JAMA 2003]
Baltimore, January 2009 18
Baltimore, January 2009 19Doyle et al, Arch Surg
2008
Baltimore, January 2009 20
“Patients who undergo liver transplantation at age 60 or above
have 1-year and 5-year survival rates similar to those of younger
patients …”
No mention that the study was of hepatitis
C positive patients
Baltimore, January 2009 21
Doyle et alEffect of age of donor
Age>60 was not statistically significant in the multivariate analysis
HR=3.03 (95%CI: 0.70-12.20) P = 0.12
“There are at least 25 other studies that have demonstrated that older donors are associated with worse outcomes in recipients with hepatitis C…”
Baltimore, January 2009 22
Importance of transparent reporting of research
Scientific manuscripts should present sufficient data so that the reader can fully evaluate the information and reach his or her own conclusions about results– Reliability and relevance
There is much evidence that this does not happen
Assessment of reliability of published articles is seriously impeded by inadequate reporting
Authors (and journals) have an obligation to ensure that research is reported adequately – i.e. transparently and completely
Baltimore, January 2009 23
Evidence of poor reporting
Poor reporting: key information is missing or ambiguous
There is considerable evidence that many published articles do not contain the necessary information – We cannot tell exactly how the research was done
Baltimore, January 2009 24
519 Randomised trials published
in December 2000
Failure to report key aspects of trial conduct:
73% Sample size calculation 55% Defined primary outcome(s)60% Whether blinded79% Method of random sequence generation82% Method of allocation concealment
[Chan & Altman Lancet 2005]
Baltimore, January 2009 25
Poor reporting is a serious problem for systematic reviews and clinical guidelines
“The biggest problem was the quality of reporting, which did not allow us to judge the important methodological items ...”
“Data reporting was poor. 15 trials met the inclusion criteria for this review but only 4 could be included as data were impossible to use in the other 11.”
(Reviews on Cochrane Library, accessed on 18 Sept 07)
Baltimore, January 2009 26
Other study types
There is most evidence for randomised trials but similar concerns apply to all types of research
Systematic reviews Phase II trials Diagnostic accuracy Epidemiology Genetic epidemiology Prognostic markers
etc
Baltimore, January 2009 27
Case-control studies
“The reporting of methods in the 408 identified papers was generally poor, with basic information about recruitment of participants often absent …”
“Poor reporting of recruitment strategies threatens the validity of reported results and reduces the generalisability of studies.”
[Lee et al. Br J Psychiatry 2007]
Baltimore, January 2009 28
Selective reporting
In addition, there is accumulating evidence of two major threats to the medical literature
Study publication bias – studies with less interesting findings are less likely to be published
Outcome reporting bias – results included within published reports are selected to favour those with statistically significant results
Baltimore, January 2009 29
Impact of poor reporting
Cumulative published evidence is misleading– Biased results– Methodological weaknesses may not be apparent
Adverse effects on– Other researchers – Clinicians– Patients
Baltimore, January 2009 30
Whose fault is poor reporting?
Poor reporting indicates a collective failure of authors, peer reviewers, and editors … on a massive scale
Researchers (authors) may not know what information to include in a report of research
Editors may not know what information should be included
What help can be given to authors?
Baltimore, January 2009 31
What can be done to improve research reports?
Research Publication Knowledge dissemination
Research Scientific writing guidanceconduct Journals’ Instructions to Authorsguidance
Baltimore, January 2009 32
What can be done to improve research reports?
Closing the gap
Research Publication Knowledge dissemination
Research Scientific writing guidanceconduct Journals’ Instructions to Authorsguidance
Reporting guidelines
Baltimore, January 2009 33
Transparency and reproducibility
All key aspects of how the study was done– Allow repetition (in principle) if desired
“Describe statistical methods with enough detail to enable a knowledgeable reader with access to the original data to verify the reported results.”
[International Committee of Medical Journal Editors]
– Same principle should extend to all study aspects– Only 49% of 80 consecutive reports accepted for
publication in Evidence-Based Medicine (2005-06) gave sufficient details of the treatment studied to allow clinicians to reproduce it [Glasziou et al 2007]
Reporting of results should not be misleading
Baltimore, January 2009 34
An early call for guidance on reporting clinical trials (1950)
“This leads one to consider if it is possible, in planning a trial, in reporting the results, or in assessing the published reports of trials, to apply criteria which must be satisfied if the analysis is to be entirely acceptable….
“A basic principle can be set up that … it is at least as important to describe the techniques employed and the conditions in which the experiment was conducted, as to give the detailed statistical analysis of results.”
[Daniels M. Scientific appraisement of new drugs in tuberculosis. Am Rev Tuberc 1950;61:751-6.]
Baltimore, January 2009 35
The CONSORT Statement for reporting RCTs[Moher et al, JAMA/Annals/Lancet 2001]
Minimum set of essential items necessary to evaluate the study
22 items that should be reported in a paper – Based on empirical evidence where possible
Also a flow diagram describing patient progress through the trial
Long explanatory paper (E&E)
www.consort-statement.org
Baltimore, January 2009 36
Goals of CONSORT
Main objective To facilitate critical appraisal and
interpretation of RCTs by providing guidance to authors about how to improve the reporting of their trials
Secondary objective To encourage and provide incentives for
researchers to conduct high-quality, unbiased randomized trials
Baltimore, January 2009 37
Impact of CONSORT
All leading general medical journals and hundreds of specialist journals support CONSORT – Not necessarily enforced
Adoption of CONSORT by journals is associated with improved reporting [Plint et al, Med J Aust 2006]
Official extensions– Cluster trials, non-inferiority and equivalence trials, harms,
non-pharmacological treatments, pragmatic trials, abstracts
Unofficial extensions– e.g. Acupuncture (STRICTA), Nonrandomised public health
interventions (TREND)
Baltimore, January 2009 38
Other guidelines
Other study types – CONSORT as a model– QUOROM (meta-analyses of RCTs) ( PRISMA)– STARD (diagnostic studies)– STROBE (observational studies)– REMARK (tumour marker prognostic studies)– …
Such guidelines are still not widely supported by medical journals or adhered to by researchers– Their potential impact is blunted
Baltimore, January 2009 39
STROBE
Baltimore, January 2009 40
Key aspects of reporting guidelines
Guidance not requirements– Journals may enforce adherence
For authors, editors, and readers
Not methodological quality“Accurate and transparent reporting is like turning the light on before you clean up a room: It doesn’t clean it for you but does tell you where the problems are.”
[Frank Davidoff, Ann Intern Med 2000]
Adherence does not guarantee a high quality study!
Baltimore, January 2009 41Perrone et al, Eur J Cancer 2008
Baltimore, January 2009 42Perrone et al, Eur J Cancer 2008
Baltimore, January 2009 43
EQUATOR: Enhancing the QUAlity and Transparency Of health Research
EQUATOR grew out of the work of CONSORT and other guidelines groups
Guidelines are available but not widely used Recognised the need to actively promote
guidelines EQUATOR Network
– Editors of general and specialty journals, researchers, guideline developers, medical writers
The goal: Better reporting, better reviewing, better editing
Baltimore, January 2009 44
www.equator-network.org
Baltimore, January 2009 45
EQUATOR Core Programme: Objectives
1. Provide resources enabling the improvement of health research reporting Website Courses
2. Monitor progress in the improvement of health research reporting
Achieving funding for such work is a major headache!
Baltimore, January 2009 46
What does the poor quality of published studies tell us about peer review?
Peer review is difficult and only partly successful
Reviewers (and editorial staff) are unable to eliminate errors in methodology and interpretation
Readers should not assume that papers published in peer reviewed journals are scientifically sound – But, many do!
Baltimore, January 2009 47
Some (partial) solutions
Monitoring/regulation– Ethics committees, data monitoring committees,
funders Trial registration Publication of protocols Availability of raw data Journal restrictions
Baltimore, January 2009 48
Reproducible research
Baltimore, January 2009 49
How much confidence can we have in published medical research?
Can’t rely on authors– Poor methodological quality– Poor reporting
Can’t rely on peer review– Fallible
In general, must rely on one’s own training, experience and judgement– Read with caution!– Be aware of what may be missing
Baltimore, January 2009 50