a novel design for decision rules based on statistical
play

A Novel Design for Decision Rules Based on Statistical Testing - PowerPoint PPT Presentation

A Novel Design for Decision Rules Based on Statistical Testing Strategies in a Definitive Go/No-Go Clinical Study Ming Zhou Bristol-Myers Squibb, New Jersey, USA - joint work with Dr. Larry Z. Shen PSI & DIA Virtual Journal Club December


  1. A Novel Design for Decision Rules Based on Statistical Testing Strategies in a Definitive Go/No-Go Clinical Study Ming Zhou Bristol-Myers Squibb, New Jersey, USA - joint work with Dr. Larry Z. Shen PSI & DIA Virtual Journal Club December 15, 2015

  2. Outline Introduction 1 Wilson CI and Hypothesis Testing 2 Separation Curve 3 Correlated Binary Data 4 Other Thoughts 5 Summary 6 Appendix 7 1 / 35

  3. Motivation Investment in new drug development is both costly and risky. Go/No-go decisions are to be made between development phases Potential approaches to go/no-go decisions ◮ Modeling and Simulation: Burman et al. (2005) and Kowalski et al. (2007) gave an overview of modelling by combining PK/PD data with clinical data and incorporating the modeling in trial simulations to improve decision-making in clinical development. ◮ Meta-Analyses ◮ Chuang-Stein et al. (2011) proposed a quantitative approach by combining the ideas of diagnostic tests and hypothesis tests for making go/no-go decisions in drug development. ◮ Nothing beats collecting and using more information to inform decisions Introduction 2 / 35

  4. A Definite Go/No-Go Study A definitive go/no-go clinical study is sometimes conducted before major investment to advance drug development into new phases. There can be many aspects for such a go/no-go study, e.g., whether certain type of adverse event could bring a potential safety risk for a Phase III program; whether a new device could be used, etc. Here we focus on binary endpoints, e.g., AE, device failure, etc. The go/no-go study is mainly for internal decision-making. Introduction 3 / 35

  5. Study Design and Hypothesis Testing Consider a single arm study where subjects are asked to test a new device. Let p represent the failure rate of the device. Select p 2 < p 1 < p 0 , such that ◮ p 0 : a failure rate at which it would not be prudent to move into the next phase. ◮ p 1 : an acceptable rate to move into next phase, where more information about the device would be collected. ◮ p 2 : ideal failure rate below which the sponsor would have total confidence to go to the next phase with the current product presentation. Very often it is unrealistic to power a go/no-go study to rule out p 2 . Introduction 4 / 35

  6. Objective Focus on the one sample test of the binomial proportion H 0 : p ≥ p 0 H 1 : p ≤ p 1 . vs Rewrite it as H 0 : p ≥ p 0 H a : p < p 0 (1) vs K 0 : p ≤ p 1 K a : p > p 1 (2) vs Examine certain situations where the difficult or ambiguous outcome might happen, e.g. we fail to reject either of H 0 and K 0 . Try to avoid ambiguous outcomes by proposing a straightforward and intuitive procedure, equipped with easy-to-interpret graphical outputs. Introduction 5 / 35

  7. A Simple Decision Rule Based on Hypothesis Testing The main goal is to allow a sponsor to make clear-cut decisions: H 0 is rejected Move to the next phase ⇒ Fail to reject H 0 Do not move to the next phase ⇒ Introduction 6 / 35

  8. Sample Size Requirement Let S n be the number of failures based on n independent Bernoulli tests. H 0 will be rejected at one-sided level α if S n ≤ c 0 , where (Fleming, 1982) � c 0 = np 0 + z α np 0 (1 − p 0 ) . (3) The sample size required for the test in (1) to have one-sided significance level α and power 1 − β is approximately � 2 � � p 0 (1 − p 0 ) + z β � p 1 (1 − p 1 ) z α n 0 = (4) . p 0 − p 1 Wilson CI and Hypothesis Testing 7 / 35

  9. Wilson Confidence Interval (CI) Let [ˆ p L ( β ) , ˆ p U ( α )] be 100 × (1 − α − β )% Wilson CI (Wilson, 1927) for p such that S n − np p ∈ [ˆ p L ( β ) , ˆ p U ( α )] ⇔ z α ≤ np (1 − p ) ≤ z 1 − β . � The Wilson CI can be used equivalently for testing (1). For simplicity of discussion, we assume that α = β , and write the CI simply as [ˆ p L , ˆ p U ]. Wilson CI and Hypothesis Testing 8 / 35

  10. Hypothesis Testing and Wilson CI For the designated sample size n 0 in (4), and H 0 , K 0 in (1), (2), If we reject H 0 , i.e., S n 0 ≤ c 0 , then ◮ ˆ p U ≤ p 0 If we fail to reject H 0 , i.e., S n 0 > c 0 , then ◮ ˆ p U > p 0 ◮ ˆ p L > p 1 (reject K 0 ). With designated sample size, we would always either reject H 0 or reject K 0 . Wilson CI and Hypothesis Testing 9 / 35

  11. Clear-Cut Decision Rules The following two decision rules are equivalent: H 0 is rejected Move to the next phase ⇒ Fail to reject H 0 Do not move to the next phase ⇒ H 0 is rejected Move to the next phase ⇒ K 0 is rejected Do not move to the next phase ⇒ Wilson CI and Hypothesis Testing 10 / 35

  12. Wilson CI and Hypothesis Testing 11 / 35

  13. Example Consider a situation where p 0 = 0 . 03 , p 1 = 0 . 01 and α = β = 0 . 05, which leads to ◮ ( n 0 , c 0 ) = (493 , 8) and ◮ a two-sided Wilson 90% confidence will be used. For each p ∈ { 0 . 005 , 0 . 01 , 0 . 02 , 0 . 03 , 0 . 04 , 0 . 05 } , 20 binomial samples from Binomial (493 , p ) are generated. Wilson CIs (two-sided 90%) are plotted and the rejection status is indicated. Wilson CI and Hypothesis Testing 12 / 35

  14. Wilson Confidence Interval With True p= 0.005 Wilson Confidence Interval With True p= 0.01 0.05 0.05 Reject H0 Reject H0 Reject K0 Reject K0 0.04 0.04 0.03 0.03 p 0 p 0 p p 0.02 0.02 ● ● ● ● ● 0.01 0.01 ● ● ● ● p 1 p 1 ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● 0.00 0.00 ● ● ● ● ● ● 5 10 15 20 5 10 15 20 Wilson Confidence Interval With True p= 0.02 Wilson Confidence Interval With True p= 0.03 Reject H0 Reject H0 Reject K0 Reject K0 0.06 0.04 ● ● ● ● 0.04 ● ● ● p 0 ● p ● p ● ● ● ● ● ● ● ● ● 0.02 ● ● ● p 0 ● ● ● ● ● 0.02 ● ● ● ● ● ● ● ● ● ● ● ● ● ● p 1 p 1 0.00 0.00 5 10 15 20 5 10 15 20 Wilson Confidence Interval With True p= 0.04 Wilson Confidence Interval With True p= 0.05 0.10 Reject H0 Reject H0 Reject K0 Reject K0 0.08 0.06 ● ● ● ● 0.06 ● ● ● ● ● ● ● ● 0.04 ● ● ● ● ● ● ● ● ● ● ● ● p p ● ● ● ● ● 0.04 ● ● ● ● ● p 0 ● ● ● 0.02 p 0 ● ● ● 0.02 p 1 p 1 0.00 0.00 5 10 15 20 5 10 15 20 Wilson CI and Hypothesis Testing 13 / 35

  15. ( n , S n , α, β ) -Separable With designated sample size and test statistic S n , we can “separate” p 0 (and anything above) from p 1 (and anything below) by ( n , S n , α, β ). By saying p 0 and p 1 are ( n , S n , α, β )-Separable, we mean True p ≤ p 1 ⇒ able to distinguish it from [ p 0 , 1] i . e ., significant evidence showing p < p 0 (type I error ≤ α ) True p ≥ p 0 ⇒ able to distinguish it from [0 , p 1 ] i . e ., significant evidence showing p > p 1 (type I error ≤ β ) With designated sample size n 0 in (4), ( p 0 , p 1 ) is ( n 0 , S n 0 , α, β )-separable. In fact, we can draw a “separation curve” for any generic ( n , S n ). Separation Curve 14 / 35

  16. 0.08 ( n , S n , 0 . 05 , 0 . 05)-separable ( p 0 , p 1 ) pairs by sample size n 0.07 0.06 design 0.05 n = 500 p 0 n = 1000 n = 2500 0.04 n = 9999 0.03 0.02 0.01 0.00 0.00 0.01 0.02 0.03 0.04 0.05 p 1 Separation Curve 15 / 35

  17. Exact Binomial vs Wilson CI ( n , S n , 0 . 05 , 0 . 05)-separable ( p 0 , p 1 ) pairs by sample size n 0.19 0.18 0.17 0.16 0.15 0.14 Method 0.13 Exact CI 0.12 Wilson CI 0.11 design p 0 0.10 50 100 0.09 250 0.08 500 0.07 750 0.06 0.05 0.04 0.03 0.02 0.01 0.00 0.00 0.01 0.02 0.03 0.04 0.05 p 1 Separation Curve 16 / 35

  18. Correlated Binary Data Outcomes from the same subjects, where multiple tests are taken. Shrinkage percentages of multiple tumors on the same subject. Defect rate of televisions from the same factory. Correlated binary data introduce intra-cluster correlation (ICC), which has been studied extensively in the literature, e.g., Fisher (1970). Designs ignoring the intra-cluster correlation can lead to inflated type I and type II error rates (Cox and Snell, 1989). Correlated Binary Data 17 / 35

  19. Correlated Binary Data Suppose there are k tests done on each of the r subjects (total number of tests done is n = rk ) Let Y ij , Y ij ′ be two of the responses from subject i . Cov ( Y ij , Y ij ′ ) = ρ p (1 − p ) , for j � = j ′ . Central Limit Theorem for Correlated Binary Data √ n (ˆ p n − p ) d → N (0 , 1) , − γ ( ρ ) p (1 − p ) � where γ ( ρ ) = 1 + ( k − 1) ρ is the variance inflation factor. Correlated Binary Data 18 / 35

  20. Effect of ICC Sample Size formula Suppose correlation is ρ , the sample size n testing (1), � 2 � p 0 (1 − p 0 ) + z β p 1 (1 − p 1 ) � � z α n ∗ = γ ( ρ ) . p 0 − p 1 Let n ∗ n = ˜ γ ( ρ ) , then ˜ n is the “effective” sample size. If ρ > 0, the maximum effective sample size is r → r n = ˜ ρ, k → ∞ . 1 − ρ + ρ k Correlated Binary Data 19 / 35

  21. An Example of ICC Effect on Sample Size Assume k = 4, α = 0 . 05 , β = 0 . 1. ρ r n 0 50 200 0.2 80 320 0.4 110 440 0.6 140 560 0.8 170 680 1 200 800 Table: Sample size for testing (1) with p 0 = 0 . 5 and p 1 = 0 . 3. Correlated Binary Data 20 / 35

Recommend


More recommend