Upload
felicia-franklin
View
212
Download
0
Tags:
Embed Size (px)
Citation preview
1
Dose-adaptive study designs offer benefits for proof-of-concept / Phase IIa clinical trials,
as well as raise issues for continued research
OUTLINE: Dose-Adaptive Designs & Examples Definition & Introduction (Jim) Frequentist Designs, including Random Walk Designs (Jim)
3+3 Design for cancer Up&Down Design Biased Coin Designs Simulations of Up&Down Design for Dental Pain Clinical Trial
Bayesian-type Designs (Inna) Continual Reassessment Method (CRM) Bayesian D-optimal Design Other related approaches Bayesian 4-parameter logistic
Case Study (adaptive cross-over design) CytelSim Software demo
Summary & Recommendations (Inna) References
Jim Bolognese & Inna Perevozskaya, Sept. 12, 2008
01/25/2006 Innovative Clinical Drug Development Conference
2
Continual Reassessment Method (CRM)
Most known Bayesian method for Phase I trials Underlying dose-response relationship is described by a 1-
parameter function family For a predefined set of doses to be studied and a binary response,
estimates dose level (MTD) that yields a particular proportion (P) of responses
CRM uses Bayes theorem with accruing data to update the distribution of MTD based on previous responses
After each patient’s response, posterior distribution of model parameter is updated; predicted probabilities of a toxic response at each dose level are updated
The dose level for next patient is selected as the one with predicted probability closest to the target level of response
Procedure stops after N patients enrolled Final estimate of MTD: dose with posterior probability closest to P
after N patients The method is designed to converge to MTD
3
Continual Reassessment Method (cont.)
Choose initial estimate
of response distribution& choose
initial dose
Obtain nextPatient’s
Observation
Update DoseResponse Model
& estimateProb. (Resp.)@ each dose
Max NReached?
Next Pt. Dose= Dose w/
Prob. (Resp.)Closest to
Target levelno
Stop.EDxx = Dose w/
Prob. (Resp.)Closest to
Target level
yes
4
Escalation With Overdose Control (EWOC) Bayesian Design
Assigns doses similarly to CRM, except for overdose control predicted probability of next assignment exceeding MTD is controlled (Bayesian
feasible design) this distinction is particularly important in oncology
Assumes a model for the dose-response curve in terms of two parameters: MTD probability of response at dose D1
EWOC updates posterior distribution of MTD based on this two-parameter model
Free software available here: http://sisyphus.emory.edu/software_ewoc.php
Reference: Z.Xu, M. Tighiouart, A. Rogatko
EWOC 2.0: Interactive Software for Dose Escalation in Cancer Phase I Clinical Trials Drug Information Journal 2007 : 41(02)
Babb, et al., 1998
5
Decision Theoretic Approaches
Similar to CRM Incorporates elements of Bayesian Decision Theory Designed to study a particular set of dose levels D1, . . ., Dk Two-parameter model for dose response with prior distributions
on the parameters Loss function minimizes asymptotic variance of dose which
yields a particular proportion of responses Posterior distribution estimates of the 2 parameters used to
derive next dose, i.e., that estimated to have desired response level
Whitehead, et al., 1995
6
Bayesian D-Optimal Sequential Design
Based on formal theory of optimal design (Atkinson and Donev, 1992)
Similar to EWOC, a constraint is added to address the ethical dilemma of avoiding extremely high doses
Uses a two parameter logistic model for dose response curve Slope & location Binary endpoint Minimum response rate fixed at 0%, maximum at 100%
Sequential procedure assigns dose at each stage which minimizes variance of posterior distribution of model parameters
Haines, et al., 2003
7
Simulated Bayesian D-Optimal Design for ED50
(http://haggis.umbc.edu/cgi-bin/dinteractive/inna1.html)
Efficacy: Percent of patients with “Response” assumed underlying distribution
Dose: 1 2 3 4 5 6 7 %Response: 30 40 55 65 75 75 75
Prior estimates: ED25 between doses 1 and 2 ED50 between doses 2 and 3
6 patients in Stage 1 for seeding purposes D-Optimal Design: 3 pts at dose 1, 2 at dose 3, 1 at dose 4 # responses: 1 1 1
24 subsequent patients (total 30 patients) entered sequentially at doses yielding minimum variance of model for ED50 estimate
Response / non-response assigned to approximate targeted %G/E distribution above
8
Simulated Bayesian D-Optimal Designfor ED50 – Results
(http://haggis.umbc.edu/cgi-bin/dinteractive/inna1.html)
Sequence of Doses assigned together with outcome(1=good/excellent response, 0=not)
Pt.No.,Dose(resp)
Pt.No.,Dose(resp)
Pt.No.,Dose(resp)
07, dose 3 (1) 15, dose 1 (0) 23, dose 1 (0)
08, dose 1 (0) 16, dose 5 (1) 24, dose 5 (0)
09, dose 3 (0) 17, dose 1 (0) 25, dose 1 (1)
10, dose 1 (1) 18, dose 5 (0) 26, dose 5 (1)
11, dose 5 (1) 19, dose 1 (1) 27, dose 1 (0)
12, dose 5 (0) 20, dose 5 (1) 28, dose 5 (0)
13, dsoe 5 (1) 21, dose 1 (0) 29, dose 1 (0)
14, dose 5 (1) 22, dose 5 (1) 30, dose 5 (1)
9
Simulated Bayesian D-Optimal Designfor ED50 – Summary
Results from a single implementation Dose: 1 2 3 4 5 6 7 assumed %Response: 30 40 55 65 75 75 75 #Responses: 4 - 2 1 8 - - #patients: 13 0 4 1 12 0 0observed %Response: 31 - 50 100 67 - -
Bayesian estimated ED50 = dose 2.3 However, few observations at other than 2 doses due to
optimal design for particular dose-response model Dose-Response curve between those two doses could be
interpolated by the underlying fitted model Should not extrapolate from model outside observed
range 2-parameter model (slope, location) forced through 0%
and 100%
10
Bayesian Design for the 4-parameter Logistic model
Underlying model:
Available doses:
Yij is (continuous) response of the j-th subject on the i-th dose, di is the vector of parameters of the distribution f
Patients are randomized in cohorts Within each cohort, fixed fraction (e.g. 25%) is allocated to
placebo, For the remaining patients within cohort, dose is picked
adaptively out of Doses are picked so that QWV (Quantile Weighted Variance)
utility function is minimized
2,( ) , ~ (0, )ij i ij ijY f d N
( ) /( , ( , , , )
(1 e )df d
minVar1
Q
q qq dfwQWV
kdd ,,1
kdd ,,1
Developed by S. Berry for CytelSim (~2006)
11
Bivariate Models:
Penalized Adaptive D-optimal Designs Addresses safety and efficacy simultaneously
Design is characterized by two dependent binary outcomes (efficacy and toxicity)
Similar to univariate model: involves dose-escalation and early stopping rules
Similar to Bayesian Sequential D-Optimal Design: Model-based approach with formal optimality criteria: “maximize the expected increment of information at
each dose” Instead of Bayesian posterior update, Maximum
Likelihood Estimates of current trial data used for next dose selection
“Penalized” design: Introduces various constraints that can be flexible to reflect ethical concerns, cost, sample size, etc.
Dragalin, 2005
12
Bayesian-type DesignsPros (+) & Cons (-)
+ Minimize observations at doses of little interest (too small or large)
+ CRM assigns doses which migrate & cluster around EDxx- little info on dose-response away from targeted dose (e.g., ED50) + can compensate by targeting 2 or 3 response levels
+ Bayesian D-optimal design efficiently estimates model-based dose-response curve (& targeted EDxx)- yields most observations at 2 dose levels to optimally fit model- model restrictive
- forced through 0% and 100% response levels - should not extrapolate response levels beyond observed
doses
13
Bayesian-type DesignsPros (+) & Cons (-)
- Subjective nature of assignment of prior (starting) distribution - could take many observations to overcome an incorrect prior
- Models underlying current methods not general enough for efficacy endpoints
- 4-parameter model needed to estimate min & max response levels
- Co-factors not included; could confound estimates+ execute designs within important co-factor levels
- Computations complex; little software available
- Difficult to explain to clients
- Not yet proven substantially better than up-and-down or t-statistic s designs when aim is estimation of dose-response curve
14
Logistics for Conduct of a Dose-Adaptive Designed Trial
Response observable reasonably quickly
Increased statistical computations / simulations to justify dose-adaptive scheme in protocol
Need on-call person to assess previous response data and generate dose for next subject For model-based dose-adaptive designs, need on-call
unblinded statistician for associated analyses OR, this could be automated via web-based interface
(increases cost) Rapid transfer of needed data
Need special packaging or unblinded pharmacist at site to package selected dose for each patient
15
Remarks (2)
Logistics of implementation more complicated than usual parallel group design Frequent data calls / brief simple analyses Close contact with sites re: dose assignments Special packaging (IVRS??) Drug Supply – needed sufficiently for many possibilities
Tolerability rule(s) can be added for downward dose-assignment if pre-specified AE criteria are encountered This has been studied in context of Bayesian dose-
adaptive designs, but not in context of up&down designs Number of placebo patients maintained as designed for
intended precision vs. that group; could be down-sized, though
16
Dose-Adaptive DesignSummary
Allocation of dose for next subject based on response(s) of previous subject(s) Random Walk designs: only last subject’s response T-statistic (frequentist) designs: all previous subjects’
responses Bayesian-type designs: all previous subjects’ responses
High potential to limit subject allocation to doses of little interest (too high / too low) Maximize information gathered from fixed N Ethical advantage over fixed randomization More attractive to patients / subjects
Inference conditional on doses assigned by design, but not overly important in early development
Requires more statistical up-front work (simulation) No pre-specified allocation schedule; requires ongoing
communication with site regarding allocation
17
Dose-Adaptive Design Summary
Bayesian-type designs preferable to estimate dose-response curve; can also estimate a dose-response quantile of interest (e.g., EDxx) or (part of?) region of increasing dose-response Complex; heavy computations
Random Walk & T-statistic Designs focus on quantile(s) of interest Easy to understand & program Consider as starting point for implementing dose-
adaptive design Let other design features guide towards other adaptive
techniques based on particular experimental situation Ongoing incomplete simulations have yet to identify
major advantage of Bayesian-type designs over RW & T, unless prior information is important to consider.
Study, comparison, & refinement of these dose-adaptive designs continues
18
Dose-adaptive study designs offer benefits for proof-of-concept / Phase IIa clinical trials,
as well as raise issues for continued research
OUTLINE: Dose-Adaptive Designs & Examples Definition & Introduction (Jim) Frequentist Designs, including Random Walk Designs (Jim)
3+3 Design for cancer Up&Down Design Biased Coin Designs Simulations of Up&Down Design for Dental Pain Clinical Trial
Bayesian-type Designs (Inna) Continual Reassessment Method (CRM) Bayesian D-optimal Design Other related approaches Bayesian 4-parameter logistic
Case Study (Bayesian and Adaptive cross-over designs) CytelSim Software demo
Summary & Recommendations (Inna) References
Jim Bolognese & Inna Perevozskaya, Sept. 12, 2008
20
Study Background
Development phase: Ib Strategic objective: generate preliminary D-R info to optimize
dose selection for Phase IIb study Caveats:
Phase Ib will be run using surrogate endpoint Future Phase IIb will be driven by clinical endpoint (chronic
symptoms) There is no formally established relationship between
surrogate and clinical endpoints dose-response curves, but… Dose selected as “sub-maximal” using surrogate endpoint D-R
curve is believed to be “sub-therapeutic” for the clinical endpoint.
21
Study objectives
(Broad) to demonstrate that a single administration of drug, compared with placebo, provides response that varies by dose
(Specific) 1. Find “sub-maximal” dose (e.g. ED75 defined as
the dose yielding 75% of the placebo-adjusted maximal response )
2. Meaningfully describe dose-response relationship
3. Demonstrate that at least one dose is significantly different from placebo
22
Study Design challenges and Adaptive design opportunity
Easy to miss informative dose range with traditional design Dose-response (D-R) can be relatively steep in the sloping part
of the D-R curve Dose-range & shape of curve = Unknown (and “unknowable”
using PK)
Dose range to explore is very wide (6 active doses potentially considered)
Logistics: Primary endpoint captured electronically within 1 day The expected subject enrolment rate is not too high
Endpoint suitable for cross-over design Following single-dose administration, 3-7 day washout is
sufficient 3-period (or even 4-period) cross-over could be reasonable
23
Bayesian Adaptive Design Description
6 active doses and placebo available Design uses frequent looks at the data and
adaptations (dose selections) are made after each IA Patients are randomized in cohorts Cohort is a small group of patients randomized between
IAs Within each cohort, fixed fraction (e.g. 25%) is allocated
to placebo For the remaining patients within cohort, dose is picked
adaptively out of D1, ….D6. Once endpoints for the whole cohort become available,
decision is made about next cohort allocation using Bayesian algorithm (QWV utility function)
24
Bayesian Adaptive Design Description (cont.)
The algorithm will try to cluster dose assignments around the “interesting” part of dose-response curve (e.g. ED75) but there will be some spread around it (i.e. not
all patients within cohort will go to the same dose).
The “target” will be moving for each cohort to be randomized will depend on the trial information accumulated to the
moment: • previous cohort’s dose allocations and responses.
25
Bayesian Adaptive Design for the 4-Parameter Logistic Model: Details
Underlying model:
Available doses:
Yij is (continuous) response of the j-th subject on the i-th dose, di is the vector of parameters of the distribution f
Patients are randomized in cohorts Within each cohort, fixed fraction (e.g. 25%) is allocated to
placebo, For the remaining patients within cohort, dose is picked
adaptively out of Doses are picked so that QWV (Quantile Weighted Variance)
utility function is minimized
2,( ) , ~ (0, )ij i ij ijY f d N
( ) /( , ( , , , )
(1 e )df d
minVar1
Q
q qq dfwQWV
kdd ,,1
kdd ,,1
Developed by S. Berry for CytelSim (~2006)
26
Implementation details of Bayesian Algorithm
Developed by Scott Berry Implemented in Cytel Simulation Bench software developed
by Cytel in collaboration with Merck Core idea: algorithm utilizes Bayesian updates of model
parameters after each cohort Components of ( , , ,) in 4-param. logistic model are
treated as random with prior distribution (usually flat) placed upon them
After each cohort’s response, the (posterior) parameter distribution is updated and model D-R is re-estimated
The algorithm utilizes Minimum Weighted Variance utility function for decision making during adaptations In our example, that translates into next cohort’s dose
assignments are picked so that the variance of the response at the current estimate of ED75 is as small as possible
27
Flexible Modeling of Dose-Response With 4-Parameter Logistic Model
0 1 2 3 4 5 6
6
8
10
12
14
16
18
20
22
24
Doses
Mea
n R
esp
on
se
Dose Response curves
ID1
ID2
ID3
ID4
ID5
5 15 3 0.5 5 15 3 0.1 5 15 1 2 5 25 5 1 -4 35 4 4
28
Bayesian Adaptive Design Allocation Example
1 2 3 4 5 6 7 8 9 100
2
4
6
8
10
12
Cohort ID
Nu
mb
er o
f S
ub
ject
s
Distribution ofSubjects
Dose 0
Dose 1
Dose 2
Dose 3
Dose 4
Dose 5
Dose 6
Sample size modeled: N=120 10 cohorts of 12
patients in parallel design setting
Different dose allocations of 12 patients in each cohort are represented by different colors (dark blue is Pbo=D0, D1 is blue, …., brown is D6).
29
Bayesian Adaptive Design Example (Cont.)
0 1 2 3 4 5 6-20
-10
0
10
20
30
40
Mea
n R
espo
nse
Subject Response Plot
0 1 2 3 4 5 6
10
20
30
40
50
30
6
26 26
13 9 10Subj
ects
Dose-response curve and Patient allocations to each dose at the end of this example trial Assume the "true" max.
effect dose (i.e., the dose to be "discovered" by the study) is "midpoint between Dose 2 and Dose 3
D-R curve captured "Dose 3" as correct start of plateau. Most patients were close to
“Dose 3” Few patients were on
plateau
30
Bayesian Algorithm Allocation rule
The algorithm clusters dose assignments around the “interesting” part of dose-response curve
With some “spread” around it (i.e. not all patients within cohort will go to the same dose).
The “target” is data-dependent and may change after each IA
31
Performance of Bayesian AD Under Various Dose-Response scenarios
8 different dose-response scenarios were studied, varying in: Magnitude of maximum
treatment effect Location of the sloping part
of the DR curve Steepness of the sloping
part of the DR curve
Allowing "true" ED75 to vary over the dose-range 0 1 2 3 4 5 6
0
2
4
6
8
10
12
14
16
18
Doses
Me
an
Re
sp
on
se
Dose Response curves
ID1
ID2
ID3
ID4
ID5
ID6
ID7
ID8
32
Performance of Bayesian AD Under Various Dose-Response scenarios (cont.)
Performance evaluated via simulations (using CytelSim Software)
Key criteria for evaluation included1. Subject allocation pattern2. Precision of picking “right dose” correctly3. Power and Type I error for detecting dose-
response 4. Precision of overall D-R estimation across all
doses (measured by MSE)
33
Subject Allocation Pattern:
ED75 centered within the dose range (Curve ID 3)
0
5
10
15
20
25
30
35
40
45
50
Nu
mb
er
of
su
bje
cts
0 1 2 3 4 5 60
10
20
3025
7.210
15 13 1317
Doses
% A
llo
ca
tio
n
0 1 2 3 4 5 6
0
2
4
6
8
10
12
14
16
18
Doses
Me
an
Re
sp
on
se
Dose Response curves
ID1
ID2
ID3
ID4
ID5
ID6
ID7
ID8
34
Subject Allocation Pattern:
ED75 shifted to the left of dose range (Curve ID 5)
0 1 2 3 4 5 6
0
2
4
6
8
10
12
14
16
18
Doses
Me
an
Re
sp
on
se
Dose Response curves
ID1
ID2
ID3
ID4
ID5
ID6
ID7
ID8
0
5
10
15
20
25
30
35
40
45
50
Nu
mb
er
of
su
bje
cts
0 1 2 3 4 5 60
10
20
3025
9.9
23
159.7 8.1 9.5
Doses
% A
llo
ca
tio
n
35
Subject Allocation Pattern:
ED75 shifted to the right of dose range (Curve ID 4)
0 1 2 3 4 5 6
0
2
4
6
8
10
12
14
16
18
Doses
Me
an
Re
sp
on
se
Dose Response curves
ID1
ID2
ID3
ID4
ID5
ID6
ID7
ID8
0
5
10
15
20
25
30
35
40
45
Nu
mb
er
of
su
bje
cts
0 1 2 3 4 5 60
10
20
3025
5.9 6.5 8.413
1922
Doses
% A
llo
ca
tio
n
36
Subject Allocation Pattern:
Completely flat dose-response (Curve ID 8)
0 1 2 3 4 5 6
0
2
4
6
8
10
12
14
16
18
Doses
Me
an
Re
sp
on
se
Dose Response curves
ID1
ID2
ID3
ID4
ID5
ID6
ID7
ID8
0
10
20
30
40
50
60
Nu
mb
er
of
su
bje
cts
Distribution of Subjects
0 1 2 3 4 5 60
10
20
30
40
25
7.2 6.8 7.4 8.1 9.4
36
Doses
% A
llo
ca
tio
n
37
Power and Type I Error for Detecting Dose-Response
D-R Curve ID “best dose” vs. pbo Test of slope (dose-resp.) 1 5.0 4.0 2 92.5 98.0 3 89.5 98.5 4 93.5 97.5 5 100.0 100.0 6 92.0 90.0 7 95.5 100.0 8 99.5 100.0
*Power for D-R curve ID 1 is Type I error
*
38
Precision of D-R estimation: Mean Squared Errors
Doses D-R
Scenarios d_0 d_1 d_2 d_3 d_4 d_5 d_6
1 4.5 2.7 2.4 2.4 2.5 2.9 4.2
2 4.4 3.7 5.0 7.0 6.4 4.6 4.7
3 3.9 5.4 7.3 6.6 3.9 3.2 3.1
4 3.1 2.9 3.1 3.9 5.7 5.1 4.1
5 4.2 6.7 5.9 3.7 3.1 3.4 4.1
6 2.6 2.5 2.5 2.7 3.2 5.0 4.9
7 3.9 3.7 4.6 7.1 6.2 3.3 4.6
8 5.5 9.5 6.5 4.1 3.1 4.0 5.8
39
MSE Efficiency Plots
0 1 2 3 4 5 6
1.4
1.6
1.8
2
2.2
2.4
2.6
2.8
3
3.2
Doses
MS
E E
ffic
ien
cy
MSE Ratio Plot
0 1 2 3 4 5 61
1.5
2
2.5
3
3.5
Doses
MS
E E
ffic
ien
cy
MSE Ratio Plot
Curve ID 3 (centered) Target dose is D3
Curve ID 5 (left-shifted) Target dose is between D2&D3
40
MSE Efficiency Plots (cont.)
0 1 2 3 4 5 61.4
1.6
1.8
2
2.2
2.4
2.6
2.8
3
3.2
3.4
Doses
MS
E E
ffic
ien
cy
MSE Ratio Plot
0 1 2 3 4 5 60.5
1
1.5
2
2.5
3
3.5
Doses
MS
E E
ffic
ien
cy
MSE Ratio Plot
Curve ID is 4 (right-shifted) target dose is D5
Curve ID is 8 (flat) target dose- NA
41
Summary of Bayesian Design Simulations
In all 7 non-flat D-R scenarios, the design maximized allocations around the “true” ED75.
In case of flat D-R, most patients were allocated to max dose and placebo with very little in between
Type I error was preserved Power to detect a dose-response is at least 90% Power to detect a significant difference between the best
dose and placebo is at least 89.5% For all scenarios, AD design was uniformly more efficient
than fixed design of the same sample size (measured by MSE ratio across doses )
42
Further Steps
Simulations have shown that Bayesian AD design may adequately address the Ib study objectives: A definitive single dose for Phase III was NOT needed General idea about D-R needed: upper/lower plateau,
sloping part
Due to absence of readily available software for crossover design, these computer simulations used N=120 in a parallel design setting It was anticipated that similar results for power and
Type 1 error could be obtained using N=30 subjects each contributing 4 measurements
43
Further Steps (cont.)
Crossover-like framework preferable to parallel design framework between/within subject variability => sample size
considerations ( 30 vs. 120) short drug half-life -> short washout period
Option 1: modify Bayesian design so that each subject can contribute multiple measurements incorporate repeated-measures in modeling and simulations Required involvement of external vendor and extra time to
complete both simulator and randomizer Option 2: consider true crossover design but change doses
adaptively Non-Bayesian approach Could be accomplished in-house within approximately the same
timeframe due to lower computational complexity Can be reduced to “standard” crossover if no dose adjustment
takes place
44
Adaptive Crossover Design Highlights
Doses explored: {D1, …, D6} of Merck-X + pbo Based on: 4 period crossover
Pbo + active doses A,B,C Values of A, B, C are subset of {D1, …, D6} and change
dynamically after each interim look (~twice weekly) Time-to-endpoint + washout is 1 week
Decision rule: pick a subset of doses {A, B, C} from {D1, …, D6} based on (non-Bayesian) utility function
Utility function: cumulative score describing proximity of each dose to target ED75 according to current estimate of D-R
D-R estimation: based on isotonic regression model
45
Adaptive X-over Algorithm details: Score function
0.0 0.2 0.4 0.6 0.8 1.0
01
03
05
07
0
Quantile, q
Sco
re, S
(q)
For each 3 dose combination, say {A,B,C}, the score is S(qA)+S(qB)+S(qC).
46
Adaptive X-over Algorithm details: Selection of 3-dose combination
Combin. 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20
Dose A 2 2 2 2 2 2 2 2 2 2 3 3 3 3 3 3 4 4 4 5
Dose B 3 3 3 3 4 4 4 5 5 6 4 4 4 5 5 6 5 5 6 6
Dose C 4 5 6 7 5 6 7 6 7 7 5 6 7 6 7 7 6 7 7 7
Score 144 135 135 135 135 135 135 126 126 126 207 207 207 198 198 198 198 198 198 189
Randomly select one of these 3 combinations and use until next adaptation.
47
Adaptive X-over Algorithm details: Interim D-R estimation at using isotonic regression
1 2 3 4 5 6 7
-4-2
02
46
810
Dose
Res
pons
e
*
*
*
*
**
*
3 best dose combinations (based on proximity to ED75):1. {3, 4, 5}2. {3, 4, 6}3. {3, 4, 7}
Algorithm randomly choose one combination out of the 3 best combinations
48
Adaptive Cross-Over Design Performance
Characteristics via Simulations Several D-R scenarios were explored Allocation pattern:
similar to Bayesian design, the algorithm allocates subjects to the neighborhood of the effective and the highest sub-effective dose levels
N=60 patients adequate to achieve ~80% or better power for “best dose” vs. placebo comparison Type I error is preserved
Caveat: effect sizes were smaller than those explored for Bayesian AD This contributed to sample size increase from N=30 (30 patients*4
obs.=120obs) for Bayesian AD to N=60 (60 patients*4 obs. =240 obs.) for the adaptive crossover design
49Dose
Re
spo
nse
0
2
4
6
8Scenario 1
1 2 3 4 5 6 7
Scenario 2 Scenario 3
Scenario 4 Scenario 5
0
2
4
6
8Scenario 6
0
2
4
6
8Scenario 7 Scenario 8 Scenario 9
1 2 3 4 5 6 7
Scenario 10 Scenario 11
1 2 3 4 5 6 7
0
2
4
6
8Scenario 12
Simulated dose-response scenarios
50
Power for testing superiority of a dose level versus placebo
Sc. Dose 1 Dose 2 Dose 3 Dose 4 Dose 5 Dose 6
1 0.02 0.02 0.02 0.02 0.02 0.02
2 0.02 0.02 0.02 0.02 0.02 0.91
3 0.02 0.02 0.02 0.02 0.84 0.91
4 0.02 0.02 0.02 0.876 0.84 0.87
5 0.02 0.02 0.728 0.87 0.81 0.85
6 0.02 0.548 0.763 0.84 0.76 0.79
7 0.74 0.612 0.56 0.85 0.50 0.79
8 0.26 0.206 0.23 0.308 0.26 0.80
9 0.25 0.221 0.23 0.314 0.72 0.83
10 0.25 0.21 0.23 0.85 0.74 0.80
11 0.23 0.20 0.69 0.85 0.72 0.79
12 0.22 0.62 0.70 0.84 0.69 0.75
51
Case Study Summary
After careful examination of many options (3 adaptive+ 3 fixed), dose-adaptive crossover design was selected
Efficiency of crossover AD over “standard” design: A “default” 4 period cross-over design will require 44 subjects to detect
the same effect sizes w/ similar power (44*4 = 176 observations) But it would explore only 3 doses and placebo
Adaptive crossover design uses 240=60*4 observations to explore 6 doses + pbo
It is 240 vs. 176 (i.e. 36% more) observations than the standard 4 period cross-over (as a cost of exploring additional 3 doses)
However, if 6 doses were explored with standard cross-over design prohibitive study duration (7 periods) would have 42*7=294 observations True efficiency of AD: 294/240=1.23 (23% more efficient)
Protocol is approved and study is underway.
52
Dose-adaptive study designs offer benefits for proof-of-concept / Phase IIa clinical trials,
as well as raise issues for continued research
OUTLINE: Dose-Adaptive Designs & Examples Definition & Introduction (Jim) Frequentist Designs, including Random Walk Designs (Jim)
3+3 Design for cancer Up&Down Design Biased Coin Designs Simulations of Up&Down Design for Dental Pain Clinical Trial
Bayesian-type Designs (Inna) Continual Reassessment Method (CRM) Bayesian D-optimal Design Other related approaches Bayesian 4-parameter logistic
Case Study (Bayesian and Adaptive cross-over designs) CytelSim Software demo
Summary & Recommendations (Inna) References
Jim Bolognese & Inna Perevozskaya, Sept. 12, 2008
53
Cytel Simulation Bench (CytelSim) Software for adaptive designs: Background and History
Developed by Cytel Statistical Software in collaboration with Merck Team
Currently exists as an in-house tool Production version planned for near future Merck (BARDS) development team:
Jim Bolognese Inna Perevozskaya Yevgen Tymofyeyev Jason Clark
54
Cytel Simulation Bench (CytelSim) Software for adaptive designs: Background and History (cont.)
Focus: model – based designs for early stage adaptive dose-ranging studies with multiple IAs
Development started shortly after 1st BARDS AD grand rounds in July 2005 Rationale for development: surging interest in innovative
design methodologies for dose-ranging studies Simulations are crucial step for implementation of such
designs Original goal: to compare performances of a Bayesian design
and up-and-down (Frequentist) design on the same platform Other methods and more sophisticated interface added later
as project needs grew
This project is still work–in-progress
55
CytelSim Underlying Methodology Overview
Applicable to early stage adaptive dose-ranging studies Provides simulations for model-based design and analysis
for such studies General set-up:
Doses of the drug and of the placebo are available for administration
Subjects are enrolled and treated in groups, known as cohorts. The size and composition of each cohort can be user-specified
and may include a fixed number of subjects assigned to placebo
Active control arm may be included as well
kdd ,,1 0d
56
CytelSim Methodology Overview (cont.)
Study Endpoints: Binary Continuous
Bayesian Methods: 4-parameter logistic model CRM
Frequentist Methods: Up & Down Design Two Up & Down Sequence Design T-test based Up & Down Design Design for Umbrella-shaped D-R curve
Note: All methods are available for both endpoints except for CRM (available for binary endpoint only)
Reviewed through examples and demo
57
T-statistic methodology highlights
Frequentist method designed to cluster allocations around target dose of interest
Can be ran for 1 or 2 targets Target is defined as difference from placebo in response or
an absolute value of response variable After each IA, T-test Based Up&Down Design assigns all
subjects in a next cohort to a single dose Dose chosen by comparing a t-statistic (based on all
previous responses at a dose) to a set of user-defined dose escalation rules: -2, -1, 0, +1, and +2 doses
More details in Example 1.Note: differences from Bayesian design are highlighted
58
Bayesian Methodology highlights
Bayesian parametric model-based method designed to cluster allocations around multiple targets of interest
Can be ran for virtually any number of targets Target is defined as a percentage of maximal response or
difference from placebo or an absolute value of a response variable
Priority of targets can be reflected through placing weights on them
After each IA, Bayesian Design assigns all subjects in a next cohort to multiple doses
Doses and allocation proportions are chosen by computing Quantile Weighted Variance (QWV) function (based on all previous responses at all doses) Doses that give the best value of QWV ( smallest variance of
response estimate) are given preference in allocation
Note: differences from T-stat design are highlighted
60
CytelSim Interface Overview
User Interface (GUI)
Computational Algorithm
Input and store simulation parameters
Output file in tabular format
Graphic summaries
Export to Excel Workbook
65
References
Rosenberger WF, Haines LM. Competing designs for phase I clinical trials: a review. Statistics in Medicine. 2002; 21:2757-2770
Haines LM, Perevozskaya I, Rosenberger WF. Bayesian optimal designs for phase I clinical trials. Biometrics. 2003; 59:561-600 (Bayesian D-optimal Design)
Stylianou M and Flournoy N. Dose Finding Using the Biased Coin Up-and-Down Design and Isotonic Regression. Biometrics. 2002; 58:171-177. (Biased Coin and Up&Down Designs)
Krams M, Lees KR, Hacke W, et al. Acute Stroke Therapy by Inhibition of Neutrophils (ASTIN), an adaptive dose-response study of UK-279,276 in Acute Ischemic Stroke. Stroke. 2003; 34:2543-2548
Dougherty TB, Porche VH, Thall PF. Maximum tolerated dose of nalmefene in patients receiving epidural fentanyl and dilute bupivacaine for postoperative analgesia. Anesthesiology, 2000; 92(4):1010-1016.
Bolognese JA. A monte carlo comparison of three up-and-down designs for dose ranging. Controlled Clinical Trials, 1983; 4:187-196. (Up&Down Designs)
Bolognese JA, Gomez HJ, Tobert JA, Rucinska EJ. The up-and-down design for dose ranging and its use in various clinical settings. ASCPT, Atlanta, 1984.
Ivanova, A., Bolognese, J. and Perevozskaya, I. (2008). Adaptive design based on t-statistic for dose-response trials. Statistics in Medicine, 27, 1581-1592. (T-statistic Design)
He W, Liu J, Bincowitz B, Quan, H. A model-based approach in the estimation of maximum tolerated dose in phase I cancer clinical trials. MRL Technical report #102, June 2004. (3+3 Design)
66
References (cont.)
Geller NL. Design of phase I and II clinical trials in cancer: a statistician’s view. Cancer Investigations 1984; 2:483–491. (3+3 design in cancer)
Durham SD, Flournoy N. Random walks for quantile estimation. In Statistical Decision Theory and Related Topics, Gupta SS, Berger JO (eds). Springer: New York, 1994; 467–476. (RWR)
O’Quigley J, Pepe M, Fisher L. Continual reassessment method: a practical design for phase I clinical trials in cancer. Biometrics 1990; 46:33–48. (CRM)
Babb J, Rogatko A, Zacks S. Cancer phase I clinical trials: efficient dose escalation with overdose control. Statistics in Medicine 1998; 17:1103 –1120.
Whitehead J, Brunier H. Bayesian decision procedures for dose determining experiments. Statistics in Medicine1995; 14:885–893 ( decision-theoretic approach)
Dragalin V., Fedorov V. Adaptive model-based designs for dose-finding studies. To appear in Journal of Statistical Planning and Inference, 2005 (bivariate models)
Reiner E, Paoletti X, O’Quigley J. Operating characteristics of the standard phase I clinical trial design. Computational Statistics and Data Analysis 1999; 30:303 –315.
Atkinson AC, Donev AN. Optimum Experimental Designs. Oxford: Clarendon, 1992.
67
QUESTIONS/ COMMENTS
/ DISCUSSION
Dose-adaptive study designs offer benefits for proof-of-concept / Phase IIa clinical trials,
as well as raise issues for continued research
68
Backups
Additional example of Bayesian Design: ASTIN trial
Select Simulation results for adaptive crossover design
69
Customized Bayesian Adaptive DesignAcute Stroke example (ASTIN Trial)
Single IV infusion of UK-279,276 (neutrophil inhibitory factor) or placebo within 6 hours of stroke symptom onset
Bayesian sequential design with real-time efficacy data capture Continuous reassessment of dose-response
Patients randomized to Placebo (at least 15%), or The 1 of 15 doses which minimizes expected variance of
ED95 response
Krams, et al. STROKE (2003)
70
Customized Bayesian Adaptive DesignAcute Stroke example
Pre-defined stopping rules: Minimum 500 pts before futility rule could be invoked Minimum 250 pts before stopping rule for efficacy could
be invoked Maximum sample size = 1300 pts
If successful, designed for extension to definitive Phase III trial
Simulations used to characterize behavior (power) of design For flat dose-response, 661 median sample size for futility
stopping w/ 5% false positive rate For 3-point efficacy benefit, 85% of simulations correctly
stopped with median sample size 595 (w/ 4 points, 97% power, median ss 320).
Krams, et al. STROKE (2003)
71
Customized Bayesian Adaptive DesignAcute Stroke example - results
•ΔSSS/Dose curve (diff from pbo) w/ 95% posterior credible intervals
•E at 2 -> efficacy threshold•F at 1 -> futility threshold
Dose of UK-279,276(mg)
74
Setup to Evaluate Performance of AD Under Various Dose-Response scenarios
AD uses 60 * 4 =240 total observations from 60 subjects.
12 different dose-response scenarios were studied, varying in: Location of the sloping part of the DR curve Steepness of the sloping part of the DR curve
Performance of AD was addressed by simulations incorporation the following features Poisson distributed subject enrollment Response generated according to the underlying true D-R,
imposing appropriate correlation structure regarding between- and within-subject variability.
Delay in response / washout period Missing observations
75
Performance of Selected AD
Key criteria for evaluation included
Subject allocation pattern Power for testing superiority of a dose level versus
placebo; Type I error Dose selection properties ( right identification of D-
R plateau ) Simultaneous confidence interval methodology for
isotonic regression is developed in the literature and can be used for dose-response curve estimation
76
Example: Mean subject allocation to dose levels and power
1 2 3 4 5 6 7
0.10
0.20
Scenario 7, dose-resp: 0, 8, 8, 8, 8, 8, 8 Power vs PLB:0, 0.75, 0.61, 0.56, 0.85, 0.49, 0.79
Dose Level
Pro
porti
on
1 2 3 4 5 6 7
0.10
0.15
0.20
0.25
Scenario 8, dose-resp: 0, 4, 4, 4, 4, 4, 8 Power vs PLB:0, 0.26, 0.21, 0.23, 0.31, 0.27, 0.8
Dose Level
Pro
porti
on
1 2 3 4 5 6 7
0.10
0.15
0.20
0.25
Scenario 9, dose-resp: 0, 4, 4, 4, 4, 8, 8 Power vs PLB:0, 0.26, 0.22, 0.23, 0.31, 0.72, 0.83
Dose Level
Pro
porti
on
1 2 3 4 5 6 7
0.10
0.20
Scenario 10, dose-resp: 0, 4, 4, 4, 8, 8, 8 Power vs PLB:0, 0.25, 0.21, 0.23, 0.85, 0.74, 0.8
Dose Level
Pro
porti
on
1 2 3 4 5 6 7
0.10
0.20
Scenario 11, dose-resp: 0, 4, 4, 8, 8, 8, 8 Power vs PLB:0, 0.23, 0.2, 0.69, 0.85, 0.72, 0.79
Dose Level
Pro
porti
on
1 2 3 4 5 6 7
0.10
0.15
0.20
0.25
Scenario 12, dose-resp: 0, 4, 8, 8, 8, 8, 8 Power vs PLB:0, 0.22, 0.61, 0.7, 0.83, 0.69, 0.75
Dose Level
Pro
porti
on
77
Example: Selecting dose level that is the onset of D-R plateau
010
2030
40True response at doses:
0, 4, 4, 4, 8, 8, 8
Dose
Pro
port
ion
(%)
none 2 3 4 5 6 7
78
Conclusions From Simulations
On average, the algorithm allocates subjects to the neighborhood of the effective and the highest sub-effective dose levels for the studied D-R scenarios. Depending on the underlying dose-response scenario,
the average number of observations on the dose with highest number of subjects ranged from 43 to 60.
Good estimation of the D-R plateau on set High power to declare effective and highest sub-effective
dose levels superior to placebo. Depending on the scenario, the power of the AD to
declare the dose with highest number of subjects significantly different from placebo ranged from 75 to 91% with most scenarios yielding power 80% or greater.