Upload
others
View
3
Download
0
Embed Size (px)
Citation preview
Fall 2007 web-based training series
Introduction
• Alex Dmitrienko (Lilly), Chair of Distance Training
Seamless adaptive designs and trial integrity issues for adaptive designs
• Jeff Maca (Novartis) and Paul Gallo (Novartis)
Presentation slides
• http://www.biopharmnet.com/doc/doc03002.html
Discussion thread
• http://biopharmnet.com/forum/viewtopic.php?t=102
Adaptive trials webinar series
Adaptive Seamless Designs for
Phase IIb/III Clinical Trials
Jeff Maca, Ph.D.
Senior Associate Director
Novartis Pharmaceuticals
Adaptive trials webinar series – January 10, 20082
Outline
Introduction and motivation of adaptive
seamless designs (ASD)
Statistical methodology for seamless designs
with examples
Logistical considerations for adaptive design
implementation
Simulations and comparisons of statistical
methods
Case study for a seamless Phase II/III design
Conclusions
Adaptive trials webinar series – January 10, 20083
Introduction and Motivation
Reducing time to market is/has/will be a top
priority in pharmaceutical development
Brings valuable medicines to patients sooner
Increases the value of the drug to the parent
company
Adaptive seamless designs can help reduce this
development time
Adaptive trials webinar series – January 10, 20084
Definitions
Seamless design
A clinical trial design which combines into a single
trial objectives which are traditionally addressed in
separate trials
Adaptive Seamless design
A seamless trial in which the final analysis will use
data from patients enrolled before and after the
adaptation (inferentially seamless)
Adaptive trials webinar series – January 10, 20085
Adaptive Seamless Designs
Primary objective – combine “dose selection”and “confirmation” into one trial
Although dose selection is most common phase IIb objective, other choices could be made, e.g. population selection
After dose selection, only change is to new enrollments (patients are generally not re-randomized)
Patients on terminated treatment groups could be followed
All data from the chosen group and comparator is used in the final analysis. Appropriate statistical methods must be used
Adaptive trials webinar series – January 10, 20086
Adaptive Seamless Designs
Dose
A
Dose B
Dose C
Placebo
Dose
A
Dose B
Dose C
Placebo
< white space >Phase II Phase III
Stage A (learning) Phase B (confirming)
Time
Traditional
Adaptive
Adaptive trials webinar series – January 10, 20087
Statistical methodology
Statistical methodology for Adaptive Seamless
Designs must account for potential biases
and statistical issues
Selection bias (multiplicity)
Multiple looks at the data (interim analysis)
Combination of data from independent stages
Adaptive trials webinar series – January 10, 20088
Simple Bonferroni adjustment
Test final hypothesis at / ntrt
Accounts for selection bias: multiplicity adjustment
Multiple looks at the data: not considered
Combination of data from stages by simple pooling
In some sense, ignores that there was an interim analysis at all
Most conservative approach, simple to implement
No other adjustments (i.e., sample size) can be made
Statistical methodology - Bonferroni
Adaptive trials webinar series – January 10, 20089
Statistical Methodology – Closed Testing
And alternative and more powerful approach is
a closed testing approach, and combination
of p-values with inverse normal method
Methodology combines:
Closed testing of hypothesis
Simes adjustment of p-values for multiplicity
Combines data (p-values) from stages via the inverse
normal method (or Fisher’s combination)
Adaptive trials webinar series – January 10, 200810
Statistical Methodology – Closed Testing
Closed test procedure
n null hypotheses H1, …, Hn
Closed test procedure
considers all intersection
hypotheses.
Hi is rejected at global level
if all hypotheses HI formed by
intersection with Hi are
rejected at local level
H1 can only be rejected
at =.05 if H12 is also
rejected at =.05
Adaptive trials webinar series – January 10, 200811
Statistical Methodology – Closed Testing
A typical study with 3 doses 3 pairwise hypotheses.
Multiplicity can be handled by adjusting p-values from
each stage using Simes procedure
iSi
S pi
Sq min S is number of elements in Hypothesis,
p(i) is the ordered P-values
Adaptive trials webinar series – January 10, 200812
Inverse Normal Method
If p1 and p2 are generated from independent data, then
will yield a N(0,1) test statistic under the null hypothesis
Note: For adaptive designs, typical value for t0 is n1/ (n1+n2)
Statistical Methodology – Closed Testing
)p1(t1)p1(t)p,p( 2
1
01
1
021C
Adaptive trials webinar series – January 10, 200813
Statistical Methodology – Example
Example: Dose finding with 3 doses + control
Stage sample sizes: n1 = 75, n2 =75
There are 3 hypotheses to be tested Hi: µi > µc for each
of the 3 treatment groups
Unadjusted pairwise p-values from the first stage:
p1,1= 0.23, p1,2 = 0.18, p1,3 = 0.08
Dose 3 selected at interim
Unadjusted p-value from second stage: p2,3 = .01
Adaptive trials webinar series – January 10, 200814
Statistical Methodology – Example
Three-way test:
q1,123 = min( 3*.08, 1.5*.18, 1*. 23)= .23
q2,123 = p2,3 = .01
C(q1,123, q2,123) = 2.17 P.value = .015
Adaptive trials webinar series – January 10, 200815
Statistical Methodology – Example
Two -way tests:
q1,13 = min( 2*.08, 1*. 23)= .16
q1,23 = min(2*.08,1*.18) = .16
q2,13 = q2,23 = p2,3 = .01
C(q1,13, q2,13) = C(q1,23, q2,23) = 2.35 P.value = .0094
Adaptive trials webinar series – January 10, 200816
Statistical Methodology – Example
Final test:
q1,3 = p1,3 = .08
q2,3 = p2,3 = .01
C(q1,13, q2,13) = C(q1,23, q2,23) = 2.64 P.value = .0042
Conclusion: Dose 3 is effective
Adaptive trials webinar series – January 10, 200817
Simulation for power comparison
To compare the methods of Bonferroni adjustment and close testing for analyzing an adaptive seamless design, the following parameters where used:
Sample sizes were n1= n2 = 75
Primary endpoint is normal, with = 12
One dose was selected for continuation
Various dose responses were assumed
20,000 reps used for simulations (error = ±.5%)
Statistical Methodology – Power
Adaptive trials webinar series – January 10, 200818
Statistical Methodology – Power
Simulation for power* comparison
Selecting 1 treatment group from 2 possible treatments
92.2%91.0%4.5, 4.5
83.2%83.1%0 , 4.5
Power Closed TestPower BonferroniDose Response
( placebo )
* Power is defined as the probability that the selected dose is confirmed
Adaptive trials webinar series – January 10, 200819
Statistical Methodology – Power
Simulation for power comparison
Selecting 1 treatment group from 3 possible treatments
92.7%90.8%4.5, 4.5, 4.5
78.9%79.4%0 ,0, 4.5
Power Closed
Test
Power
Bonferroni
Dose Response
( placebo )
Adaptive trials webinar series – January 10, 200820
Considerations for Seamless Designs
With the added flexibility of seamless designs,
comes added complexity.
Careful consideration should be given to the feasibility
for a seamless design for the project.
Not all projects can use seamless development
Even if two programs can use seamless development,
one might be better suited than the other
Many characteristics add or subtract to the feasibility
Adaptive trials webinar series – January 10, 200821
Considerations for Seamless Designs
Enrollment vs. Endpoint
The length of time needed to make a decision relative
to the time of enrollment must be small
Otherwise enrollment must be paused
Endpoint must be well known and accepted
If the goal of Phase II is to determine the endpoint for
registration, seamless development would be difficult
If surrogate marker will be used for dose selection, it
must be accepted, validated and well understood
Adaptive trials webinar series – January 10, 200822
Considerations for Seamless Designs
Clinical Development Time
There will usually be two pivotal trials for registration
Entire program must be completed in shorter timelines,
not just the adaptive trial
Adaptive trials webinar series – January 10, 200823
Considerations for Seamless Designs
Logistical considerations
Helpful if final product is available for adaptive trial
(otherwise bioequivalence study is needed)
Decision process, and personnel must be carefully
planned and pre-specified
Adaptive trials webinar series – January 10, 200824
Considerations for Seamless Designs
Novel drug or indication
Decision process which will be overly complicated could
be an issue with an external board
If there are a lot of unknown issues with the indication
or drug, a separate phase II trial would be better
However, getting a novel drug to patients sooner
increases the benefit of seamless development
Adaptive trials webinar series – January 10, 200825
Statistical Considerations
Bias of treatment estimates
Selection of treatments and combing data from both stages will lead to a positive bias in the treatment effect
Simulations can be run to quantify this bias
The data from the second stage could be used to produce an unbiased estimate of the treatment effect
Bias is larger when more treatment groups are present in the first stage
Bias is increased if the first stage comprises a larger part of the final combined data (i.e. as n1/n2 increases)
Adaptive trials webinar series – January 10, 200826
Choosing sample sizes
There are two sample sizes to consider for a seamless design, n1, n2
If t is the number of treatments and one dose is selected, the total same size N is:
N = t*n1 + 2*n2
The larger n1, the better job of choosing the “right” dose. However, this makes the total much larger.
Power can be determined by simulation, and is also a function of the (unknown) dose response
Statistical Considerations
Adaptive trials webinar series – January 10, 200827
Statistical Considerations
Possible dose responses
Various dose responses should be examined to
determine the effect on the study
Power
Selection probabilities
Adaptive trials webinar series – January 10, 200828
Phase II/III adaptive design: Case study
Introduction
Adaptive seamless design (ASD) to confirm
dose selection, to support registration and label
claims
Indication is a chronic disease
Study will provide pivotal confirmation of
efficacy, safety and tolerability of selected
doses
A second pivotal study will be used for
registration
Adaptive trials webinar series – January 10, 200829
Novartis dose 1
Novartis dose 2
Novartis dose 3
Novartis dose 4
Placebo
Active control 1
Active control 2
Novartis dose A
Novartis dose B
Placebo
Active control 2
Dose Ranging
2 weeks
Screening Interim
Analysis
Efficacy and Safety
26 weeks
STAGE 2 (phase III)STAGE 1 (phase IIb)
Ongoing treatment
Phase II/III adaptive design : Case study
Final
Analysis
Independent
Dose
Selection
Adaptive trials webinar series – January 10, 200830
Phase II/III adaptive design: Case study
Primary endpoint
Continuous variable – measured after 12 weeks
Comparison with placebo for superiority
Key secondary endpoint
Continuous variable - measured after 12 weeks
Comparison with active control for non-inferiority
Important secondary endpoint
Continuous variable – subjective QOL
Comparison to placebo for superiority
Multiple additional secondary endpoints
Adaptive trials webinar series – January 10, 200831
Phase II/III adaptive design: Case study
Objective of interim analysis
To investigate four doses of new treatment versus
placebo and active controls with respect to primary
endpoint after 2 weeks of treatment (early read-out)
To investigate four doses of new treatment versus
placebo and active controls with respect to
cumulative selected safety data (key AEs)
Independent external DMC to select 2 adjacent
doses based on pre-defined guidelines
Adaptive trials webinar series – January 10, 200832
Phase II/III adaptive design: Case study
Sample sizes for stage 1 and 2 based on
extensive simulation work
Stage 1 (7 arms): 115 patients/arm (805 total)
Stage 2 (4 arms): 285 patients/arm (1140 total)
Requirements for the simulation
Minimum clinically important difference (MCID),
standard deviation (SD),
Correlation between 2 and 12 week endpoints
Estimated dose response for treatment and controls
Interim analysis decision guidelines
Adaptive trials webinar series – January 10, 200833
Phase II/III adaptive design: Case study
Output from simulation
Selection probabilities for each of the treatment
dose pairs
Power conditional on the dose pair selected
Overall power that one or both selected doses
can be confirmed
Adaptive trials webinar series – January 10, 200834
Phase II/III adaptive design: Case study
Statistical methodology
A multiplicity correction of /4 will be used for
the final analysis
Conservative adjustment since only 2 treatment
doses will be used in final analysis
Similar power to inverse normal/closed testing
methodology
Greater flexibility for the testing of primary and
secondary endpoints (Allows sequential testing)
Procedure will control family-wise type I error
rate for primary and secondary endpoints
Adaptive trials webinar series – January 10, 200835
Phase II/III adaptive design: Case study
DMC guidelines
Numerical values given (not inferential)
Threshold was defined as the maximum of:
Minimum Clinically Important Difference (MCID)
Primary endpoint for both active controls (2 values)
versus placebo
The dose selected ill be the lowest dose which
exceeds this threshold, and the next highest
dose
Adaptive trials webinar series – January 10, 200836
Phase II/III adaptive design: Case study
DMC guidelines
The dose selected will be the lowest dose
which exceeds this threshold, and the next
highest dose
If the highest dose is the only dose to exceed the
threshold, then the next lowest dose will be used
If no doses exceed the threshold but are close, two
highest doses will be used
DMC can weigh any safety signal versus efficacy in
selecting a dose. DMC has discretion to deviate in
the case of unexpected results
Adaptive trials webinar series – January 10, 200837
Phase II/III adaptive design: Case study
FDA experience
“Special Protocol Assessment” used to discuss
protocol, including briefing book and draft
protocol
Face-to-face meeting to discuss ASD and other
project related issues
Protecting trial integrity: who produces interim report
DMC composition and knowledge flow (see next
presentation)
Basis for dose selection
Statistical methodology was not the greatest concern
Adaptive trials webinar series – January 10, 200838
Conclusions
Adaptive seamless designs have an ability to improve
the development process by reducing timelines for
approval
Statistical methods are available to account for adaptive
trial designs
Extra planning is necessary to implement an adaptive
seamless design protocol
Benefits should be carefully weighed against the
challenges of such designs before implementation
Adaptive trials webinar series – January 10, 200839
Primary PhRMA references
PhRMA White Paper sections:
Maca J, Bhattacharya S, Dragalin V, Gallo P, and
Krams M. Adaptive Seamless Phase II/III Designs
– Background, Operational Aspects, and
Examples. Drug Information Journal. 2006; 40(4):
463-473.
Gallo P. Confidentiality and trial integrity issues for
adaptive designs. Drug Information Journal. 2006;
40(4): 445-450.
Adaptive trials webinar series – January 10, 200840
ReferencesDragalin V. Adaptive designs: terminology and classification. Drug Inf J. 2006 (to appear).
Quinlan JA, Krams M. Implementing adaptive designs: logistical and operational considerations. Drug Inf J. 2006 (to appear).
Bechhofer RE, Kiefer J, Sobel M. Sequential Identification and Ranking Problems.Chicago: University of Chicago Press;1968.
Paulson E. A selection procedure for selecting the population with the largest mean from k normal populations. Ann Math Stat. 1964;35:174-180.
Thall PF, Simon R, Ellenberg SS. A two-stage design for choosing among several experimental treatments and a control in clinical trials. Biometrics 1989;45:537-547.
Schaid DJ, Wieand S, Therneau TM. Optimal two stage screening designs for survival comparisons. Biometrika 1990;77:659-663.
Stallard N, Todd S. Sequential designs for phase III clinical trials incorporating treatment selection. Stat Med. 2003;22:689-703.
Follman DA, Proschan MA, Geller NL. Monitoring pairwise comparisons in multi-armed clinical trials. Biometrics 1994;50:325-336.
Hellmich M. Monitoring clinical trials with multiple arms. Biometrics 2001;57:892-898.
Adaptive trials webinar series – January 10, 200841
References
Bischoff W, Miller F. Adaptive two-stage test procedures to find the best treatment in clinical trials. Biometrika 2005;92:197-212.
Todd S, Stallard N. A new clinical trial design combining Phases 2 and 3: sequential designs with treatment selection and a change of endpoint. Drug Inf J. 2005;39:109-118.
Bauer P, Köhne K. Evaluation of experiments with adaptive interim analyses. Biometrics 1994;50:1029-1041.
Bauer P, Kieser M. Combining different phases in the development of medical treatments within a single trial. Stat Med. 1999;18:1833-1848.
Brannath W, Posch M, Bauer P. Recursive combination tests. J Am Stat Assoc.2002;97:236-244.
Müller HH, Schäfer H. Adaptive group sequential designs for clinical trials: combining the advantages of adaptive and classical group sequential approaches. Biometrics 2001;57:886-819.
Liu Q, Pledger GW. Phase 2 and 3 combination designs to accelerate drug development. J Am Stat Assoc. 2005;100:493-502.
Posch M, Koenig F, Brannath W, Dunger-Baldauf C, Bauer P. Testing and estimation in flexible group sequential designs with adaptive treatment selection. Stat Med. 2005;24:3697-3714.
Bauer P, Einfalt J. Application of adaptive designs – a review. Biometrical J. 2006;48:1:14.
Adaptive trials webinar series – January 10, 200842
References
Inoue LYT, Thall PF, Berry DA. Seamlessly expanding a randomized phase II trial to phase III. Biometrics 2002;58:823-831.
Berry, DA, Müller P, Grieve AP, Smith M, Parke T, Blazek R, Mitchard N, Krams M. Adaptive Bayesian designs for dose-ranging drug trials. In Case Studies in Bayesian Statistics V. Lecture Notes in Statist. Springer: New York;2002;162:99-181.
Coburger S, Wassmer G. Sample size reassessment in adaptive clinical trials using a bias corrected estimate. Biometrical J. 2003;45:812-825.
Brannath W, König F, Bauer P. Improved repeated confidence bounds in trials with a maximal goal. Biometrical J. 2003;45:311-324.
Sampson AR, Sill MW. Drop-the-losers design: normal case. Biometrical J. 2005;47:257-281.
Stallard N, Todd S. Point estimates and confidence regions for sequential trials involving selection. J. Statist. Plan. Inference 2005;135:402-419.
US Food and Drug Administration. Guidance for Clinical Trial Sponsors. Establishment and Operation of Clinical Trial Data Monitoring Committees. 2006; Rockville MD: FDA. http://www.fda.gov/cber/qdlns/clintrialdmc.htm.
US Food and Drug Administration. Challenge and Opportunity on the Critical Path to New Medicinal Products. 2006.http://www.fda.gov/oc/initiatives/criticalpath/whitepaper.html.
Adaptive trials webinar series
Operational Issues and Trial
Integrity Concerns for Adaptive
Trials
Paul Gallo, Ph.D.
Biometrical Fellow
Novartis Pharmaceuticals
Adaptive trials webinar series – January 10, 200844
Outline
Motivations, opportunities, and challenges
Logistics and feasibility issues
Interim monitoring and confidentiality issues
review of current conventions
monitoring processes for adaptive designs
information conveyed by adaptive designs
Investigating homogeneity
Adaptive trials webinar series – January 10, 200845
General motivation
The greater flexibility offered within the adaptive
design framework has the potential to translate into
more ethical treatment of patients within trials
(possibly including the use of fewer patients), more
efficient drug development, and better focusing of
available resources.
The potential appeal of adaptive designs is
understandable, and motivates the current high
level of interest in this topic.
Adaptive trials webinar series – January 10, 200846
Cautions
But, being too eager, and proceeding without all
relevant issues being fully considered, is not
advisable either.
The question should be:
“What is the most appropriate (e.g., ethical, efficient)
means at hand to address the research questions of
importance?”
rather than:
“How can adaptive designs be integrated into our
program at all costs?”
Adaptive trials webinar series – January 10, 200847
Challenges
Clearly, there will be many challenges to be
addressed or overcome before adaptive
designs become more widely utilized.
Statistical
Logistic
Procedural / regulatory
Adaptive trials webinar series – January 10, 200848
General considerations
Like any new technology with challenges, some
resistance is to be expected.
Closer scrutiny is natural, and constructive.
But we should not make “the perfect be the
enemy of the good ”.
Can we address the challenges to a sufficient
extent so that in particular situations the
advantages outweigh the drawbacks?
Adaptive trials webinar series – January 10, 200849
Opportunities
Early-phase trials may in the short term be the
most favorable arena for wider-scale
implementation of adaptive designs.
More uncertainties, and thus more opportunity for
considering adaptation
Lesser regulatory concerns
Lower-risk opportunities to gain experience with ADs
to learn, solve operational problems, and set the
stage for more important applications.
Adaptive trials webinar series – January 10, 200850
Feasibility issues
Endpoint follow-up time vs recruitment speed
Shorter read-out time is generally favorable to
adaptive designs.
Surrogates / early predictors can have a role.
Timely data collection is important, as well as
efficient analysis and decision-making
processes.
Electronic Data Capture should be helpful.
Adaptive trials webinar series – January 10, 200851
Data quality
All else being equal, cleaner is better.
But the usual trade-off exists:
cleaner takes longer, and results in less data being
available for decisions
lack of data is a source of noise also!
There is no requirement that data must be fully
cleaned for adaptive designs.
Details of data quality requirements should be
considered on a case-by-case basis.
Adaptive trials webinar series – January 10, 200852
Monitoring / confidentiality issues
Issues relating to
monitoring of accruing data
restriction of knowledge of interim results
and the processes of data review, decision-
making and implementation
are likely to be critical in determining the extent
and shaping the nature of adaptive design
utilization in clinical trials.
Adaptive trials webinar series – January 10, 200853
Current monitoring conventions
Monitoring of accruing data is of course a common
feature in clinical trials. Most frequently for:
safety monitoring
formal group sequential plan allowing stopping for
efficacy
lack of effect / futility judgments.
Current procedures and conventions governing
monitoring are a sensible starting point for
addressing similar issues in adaptive trials.
Adaptive trials webinar series – January 10, 200854
Current monitoring conventions
As described in the FDA DMC guidance
(2006): Unblinded data and comparative
interim results should not be accessible to
trial personnel, sponsor, investigators.
Access to interim results diminishes the
ability of trial personnel to manage the trial in
a manner which is (and which will be seen
by interested parties to be) completely
objective.
Adaptive trials webinar series – January 10, 200855
Current monitoring conventions
Knowledge of interim results could introduce
subtle, unknown biases into the trial,
perhaps causing slight changes in
characteristics of patients recruited,
administration of the intervention, endpoint
assessments, etc.
Changes in “investigator enthusiasm”?
The equipoise argument: knowledge of
interim results violates equipoise.
Adaptive trials webinar series – January 10, 200856
Current conventions - sponsor
FDA (2006): “Sponsor exposure to
unblinded interim data . . . can present
substantial risk to the integrity of the trial.”
Risks include lack of objectivity in trial
management; further unblinding, even if
inadvertent; SEC requirements and fiduciary
responsibilities, etc.
Sponsor is thus typically not involved in
monitoring of confirmatory trials.
Adaptive trials webinar series – January 10, 200857
A conflict ?
Consider as a motivating example a long-term
seamless phase II/III trial with dosage selection.
In some trials, the adaptation decision:
is more traditionally a sponsor responsibility; is
complex, with sponsor perspective potentially
relevant; and can involve important sponsor
interests.
But practices in more familiar monitoring settings
hold that sponsors not have access to interim data.
Adaptive trials webinar series – January 10, 200858
Sponsor concerns
Sponsor concern: unanticipated complexities
that might not fit a pre-specified algorithm.
e.g.: unusual dose-response pattern, unexpected risk-
benefit profile, compliance issues, evolving clinical
practices, etc.
Can sponsors be comfortable without some
participation in the decision process, or at least the
ability to ratify certain important decisions?
Can we find operational models that reasonably
satisfy the competing concerns?
Adaptive trials webinar series – January 10, 200859
Issues for adaptive designs
I. Adaptive designs will certainly require review of
accruing data, and for additional purposes
beyond those in more familiar monitoring
settings. Questions:
Who will be involved in the analysis, review, and
decision-making processes?
Will operational models differ from those we’ve
become familiar with?
Adaptive trials webinar series – January 10, 200860
Issues for adaptive designs
I. Questions for monitoring models (continued):
Will sponsor perspective and input be relevant or
necessary for some types of adaptations?
Can this be achieved without undermining trial
integrity?
Can sponsors accept and trust decisions made
confidentially by external DMCs in long-term trials
/ projects with important business implications
(e.g., seamless Phase II / III)?
Adaptive trials webinar series – January 10, 200861
Issues for adaptive designs
II. An important distinction versus common monitoring
situations: the results will be used to implement
adaptation(s) which will govern some aspect of
the conduct of the remainder of the trial.
Can observers infer from viewing the actions taken
information about the results which might be perceived
to rise to an unacceptable level?
Adaptive trials webinar series – January 10, 200862
Analysis / review / decision process
Concerns about confidentiality to ensure objective
trial management, and potential bias from broad
knowledge of interim results, should be no less
relevant for adaptive designs than in other settings.
The key principles to adhere to would seem to be:
separation / independence of the DMC from other
trial activities
restriction of knowledge of interim comparative
results or unblinded data.
Adaptive trials webinar series – January 10, 200863
Analysis / review / decision process
Adaptive design trials may utilize a single monitoring
board for adaptations and other responsibilities
(e.g., safety); or else a separate board may be
considered for the adaptation decisions.
DMCs in adaptive design trials may require
additional expertise not traditionally represented on
DMCs; perhaps to monitor the adaptation algorithm,
or to make the type of decision called for in the
adaptation plan (e.g., dose selection).
Adaptive trials webinar series – January 10, 200864
Sponsor participation proposal
Proposed model for sponsor involvement
(PhRMA AD Working Group)
A sponsor role in the process for making certain
types of decisions should require:
a clear rationale for the involvement, specific to the
case at hand, based on the complex nature of the
decision and its implications;
individuals properly ‘distanced’ from trial
operations;
Adaptive trials webinar series – January 10, 200865
Sponsor participation
Model for sponsor involvement (continued):
clear understanding by all involved of the issues
and risks to the trial, documentation of the
processes followed, and secure firewalls /
procedures in place;
sponsor exposure to results is “minimal” for the
needed decision – smallest number of individuals,
only at the adaptation point, only the relevant data,
etc. (e.g., unlike an independent DMC with whom
they may be interacting, which may have a broader
ongoing role).
Adaptive trials webinar series – January 10, 200866
Sponsor involvement
Data flow processes and procedures for
restricting access from those involved in the
trial will need to be documented and
described in detail, typically in the DMC
Charter.
The burden will necessarily fall on the
sponsor to make a strong case that effective
procedures and safeguards are in place, and
are followed.
Adaptive trials webinar series – January 10, 200867
Sponsor participation
This is not a one size fits all issue! – the
circumstances should determine the needed level of
involvement.
If there is a pre-specified adaptation algorithm that
all agree will reasonably suffice, the sponsor can
and should forgo this involvement.
If there is too much potential complexity, so that
the first-stage data would need a more extensive
level of access and examination, then this would
probably argue against using the adaptive design.
Adaptive trials webinar series – January 10, 200868
Information conveyed to observers
Adaptive designs may lead to changes in a trial
which will be apparent to some extent - sample
size, randomization allocation, population, dosage,
treatment arm selection, etc., etc. - and can thus be
viewed as providing some information to observers
about the results which led to those changes.
Considering the concerns which are the basis for
the confidentiality conventions: can we distinguish
between types and amounts of information, and
how risky they would be in this regard?
Adaptive trials webinar series – January 10, 200869
Information conveyed to observers
Note: conventional monitoring is not immune
from this issue.
It has never been the case that no information
can be inferred from monitoring; i.e., all
monitoring has some potential action
thresholds, and lack of action usually implies
that such thresholds have not been reached.
Adaptive trials webinar series – January 10, 200870
Example – Triangular test
Design:
Normal data, 2 group comparison
Study designed to detect = 0.15
4 equally-spaced analyses
will require about 2276 patients.
Adaptive trials webinar series – January 10, 200871
Example – Triangular test
-40
-20
0
20
40
60
80
0 100 200 300 400 500 600 700 800
V
Z
‘Christmas tree’ boundary
Adaptive trials webinar series – January 10, 200872
Example – Triangular test
= Z / V
Continuation beyond the 3rd look would imply
(barring over-ruling of the boundary) that the point
estimate is between 0.076 and 0.106.
Doesn’t that convey quite a bit of information about
the interim results?
^
Adaptive trials webinar series – January 10, 200873
Information apparent to observers
In conventional group sequential design practice,
this issue seems not to be commonly perceived to
compromise trials nor to discourage monitoring.
Presumably, it’s viewed that reasonable balance is
struck between the objectives and benefits of the
monitoring and any slight potential for risk to the
trial, with appropriate and feasible safeguards in
place to minimize that risk.
The same type of standard should make sense for
adaptive designs.
Adaptive trials webinar series – January 10, 200874
Limiting the concern
In some cases we may have opportunities to lessen
this concern by withholding certain details of the
strategy from the protocol, and placing them in
another document of more limited circulation.
For example, if some type of selection decision will
be made based upon predictive probabilities, do full
details and thresholds need to be described in the
protocol?
Adaptive trials webinar series – January 10, 200875
Proposal – selection decisions
(PhRMA AD Working Group):
Selection decisions (choice of dose,
subgroup, etc. as in a seamless design)
generally do NOT convey an amount of
information that would be considered to
compromise or influence the trial, as long
as the specific numerical results on which
the decisions were based remain
confidential.
Adaptive trials webinar series – January 10, 200876
Information apparent to observers
Consider the alternative -
In a seamless Phase II / III design, we might instead
have run a conventional separate-phase program.
Phase II results would be widely known (what about
equipoise ??)
In this sense, maybe the adaptive design offers a
further advantage relative to the traditional
paradigm?
Adaptive trials webinar series – January 10, 200877
Algorithmic changes
More problematic - changes based in an algorithmic
manner on interim treatment effect estimates in
effect provide knowledge of those estimates to
anyone who knows the algorithm and the change.
Most typical example - certain approaches to
sample size re-estimation:
SSnew = f (interim treatment effect estimate)
=> estimate = f -1 (SSnew)
Adaptive trials webinar series – January 10, 200878
Mitigating the concerns
Perhaps the adaptation can be made based upon a
combination of factors in order to mask the observed
treatment effect.
e.g., SS re-estimation using the treatment effect, the
observed variance, and external information.
If possible, “discretize” the potential actions, i.e., a
small number of potential actions corresponds to
intervals / ranges of the treatment effects.
We may at times try to quantify that the knowledge
which can be inferred is comparable to that of
accepted group sequential plans.
Adaptive trials webinar series – January 10, 200879
Heterogeneity concerns
In adaptive trials, might some aspect of the
adaptation process change the conduct of the trial,
and limit the interpretability of the overall results?
CHMP (2007): “ . . . whenever trials are planned to
incorporate design modifications based on the results
of an interim analysis, the applicant must pre-plan
methods to ensure that results from different stages
of the trial can be justifiably combined.”
How should this investigation be implemented, and
what is its role in the overall interpretation?
Adaptive trials webinar series – January 10, 200880
Investigating homogeneity
The concern is understandable, but there may be
challenges in implementation:
Formal statistical tests for interaction across stages
will not have good operating characteristics, and
may be prone to errors of both types (signals of
change due solely to chance, underpowering for
true important effects).
Setting formal heterogeneity standards for adaptive
trials seems problematic, because we do not in
general have standards for investigating this issue
in other trials.
Adaptive trials webinar series – January 10, 200881
Investigating homogeneity
Implementation challenges (continued):
This issue is confounded with other mechanisms that
could lead to within-trial change which might be totally
unrelated to its adaptive nature (e.g., investigator
“learning curve”, geographic or demographic patient
population shifts, non-constant hazard ratios, etc.)
It might be challenging to properly and
unambiguously define the ‘stages’ being compared.
Dialog and research is ongoing (e.g., EMEA/EFPIA
Workshop, Dec. 2007).
Adaptive trials webinar series – January 10, 200882
Summary
We should not aim to broadly undo established
monitoring conventions, but rather to fine-tune them
to achieve their sound underlying principles.
To justify sponsor participation in monitoring, provide
convincing rationale and “minimize” this involvement,
and enforce strict control of information.
Some types of adaptations convey limited information
for which it seems difficult to envision how the trial
might be compromised.
Others convey more information, but perhaps we can
implement extra steps to mask this.
Adaptive trials webinar series – January 10, 200883
ReferencesCommittee for Medicinal Products for Human Use. Guideline on Data Monitoring
Committees. London: EMEA; 2006.
Committee for Medicinal Products for Human Use. Reflection paper on
methodological issues in confirmatory clinical trials planned with an adaptive design.
London: EMEA; 2007.
DeMets DL, Furberg CD, Friedman LM (eds.). Data Monitoring in Clinical Trials: A
Case Studies Approach. Springer; 2006.
Ellenberg SE, Fleming TR, DeMets DL. Data Monitoring Committees in Clinical Trials:
A Practical Perspective. Chichester: Wiley; 2002.
Gallo P. Confidentiality and trial integrity issues for adaptive designs. Drug
Information Journal 2006; 40(4): 445-450.
Quinlan JA and Krams M. Implementing adaptive designs: logistical and operational
considerations. Drug Information Journal 2006; 40(4): 437-444.
US Food and Drug Administration. Guidance for Clinical Trial Sponsors on the
Establishment and Operation of Data Monitoring Committees. Rockville MD: FDA;
2006.