1 of 56

A Rigorous Framework for Trial Design via Simulation

September 8, 2021

Michael Sklar

Stein Postdoctoral Fellow at Stanford

https://mikesklar.github.io/thesis/

2 of 56

The Innovation Process ?

Clinician

Statistician

New Design

Pharma Runs Trial

3 of 56

The Innovation Process

New Design

Pharma Runs Trial

Journal Submission

4 of 56

The Innovation Process

New Design

Pharma Runs Trial

Journal Submission

Persuade Pharma Decisionmakers

5 of 56

The Innovation Process

New Design

Pharma Runs Trial

Journal Submission

Persuade Pharma Decisionmakers

High Reward/

Low Execution Risk

6 of 56

The Innovation Process

New Design

Pharma Runs Trial

Journal Submission

Persuade Pharma Decisionmakers

High Reward/

Low Execution Risk

Low Risk of Slow Processing or Denial

7 of 56

The Innovation Process

New Design

Pharma Runs Trial

Journal Submission

Persuade Pharma Decisionmakers

Persuade FDA Decisionmakers

High Reward/

Low Execution Risk

Low Risk of Slow Processing or Denial

8 of 56

The Innovation Process

New Design

Journal Submission

Persuade Pharma Decisionmakers

High Reward/

Low Execution Risk

Pass FDA validation and negotiations

Type I Error Proof

Low Risk of Slow Processing or Denial

Pharma Runs Trial

Persuade FDA Decisionmakers

9 of 56

The Innovation Process

New Design

Journal Submission

Persuade Pharma Decisionmakers

High Reward/

Low Execution Risk

Pass FDA validation and negotiations

Type I Error Proof

Low Risk of Slow Processing or Denial

Pharma Runs Trial

Persuade FDA Decisionmakers

10 of 56

The Innovation Process

New Design

Journal Submission

Persuade Pharma Decisionmakers

High Reward/

Low Execution Risk

Pass FDA validation and negotiations

Type I Error Proof

Low Risk of Slow Processing or Denial

Pharma Runs Trial

Persuade FDA Decisionmakers

11 of 56

Simulation slices out pain points

New Design

Journal Submission

Persuade Pharma Decisionmakers

High Reward/

Low Execution Risk

Pass FDA validation and negotiations

Type I Error Proof

Low Risk of Slow Processing or Denial

Pharma Runs Trial

Persuade FDA

Decisionmakers

12 of 56

(Lifted from FDA website – John Scott, 2018)

13 of 56

14 of 56

15 of 56

A Rigorous Framework for Type I Error Control in Complex Trial Design

  • Idea: Provably control Type I error by filling in the gaps between simulation grid-points

  • Can this speed up validation?

16 of 56

What could automated validation do?

  • Reduce burden on regulators
  • Speed up validation for pharma & consumers
  • Increase predictability for designers

  • Proof can answer concerns about powerful “black box” optimizations moving Type I Error to places where it isn’t checked

17 of 56

Challenges

  • Massive computational power
    • [Example uses 10^10 simulations for 2 unknown parameters]
    • Difficulty quickly increases with complexity and # of parameters

  • Regulatory agreement on the model class for simulations

  • Well-behaved model for the data (exponential family)
    • Includes Gaussian, binomial, exponential, gamma, some Weibull models
    • (Censoring and adaptive sampling are OK)

  • Justification for focusing on a compact part of null hypothesis space

18 of 56

Roadmap for the Talk:

  • How to prove Type I Error control over continuous space with simulation
  • Examples
  • Challenges, recommendations, and additional features
    • Tuning critical values
    • Re-design and Platform trials
  • Questions
  • Discussion on possibilities for software implementation

  • Further math details

19 of 56

Consider the most basic test

  •  

20 of 56

Consider the most basic test

  •  

21 of 56

Zoom in on Type I Error near the Ho boundary

22 of 56

Assume we know the exact Type I Error only at a few points…

23 of 56

Assume we know the exact Type I Error only at a few points…

24 of 56

Assume we know the true derivative of Type I Error at those points too

25 of 56

First-order approximation is close but not conservative

26 of 56

Taylor’s Theorem Describes the Error

27 of 56

But using a worst-case bound on the second derivative, we get a conservative approximation

28 of 56

Double the number of simulation points

29 of 56

Double the number of simulation points

30 of 56

Double the number of simulation points

31 of 56

Double the number of simulation points

32 of 56

Monte Carlo simulations leaves uncertainty

33 of 56

Monte Carlo simulations leaves uncertainty

34 of 56

Key Steps:

  • Find a general upper bound to the second derivative
  • Construct confidence intervals with the Monte Carlo simulation for the point estimate of Type I Error and the derivative
    • Give each interval a confidence of 1 - 𝛿 / 2
  • Maximize our Taylor expansion upper bound over these confidence intervals

Result: A (pointwise) 1 - 𝛿 upper confidence bound to the true Type I Error function

35 of 56

Example: FWER for Two Arms

What if you did two independent z-tests?

Here is a display of the true FWER, as a function of the parameters

36 of 56

Our 99% confidence upper bound

37 of 56

Adaptive Trial Example: Thompson Sampling

  • Two arms
  • Bernoulli (𝛉i) outcomes
  • Ho : 𝛉i < .6
  • N = 100
  • Beta(1,1) prior

  • Reject arm i at the end if posterior�P(𝛉 i > .6) > 95%

With a cluster:�~800,000 Monte Carlo samples per point

~16,000 grid points

38 of 56

Setting: Data is Exponential Family

    • Includes your favorite R.V.’s : Gaussian, Binomial, Exponential, Gamma, Weibull, …

    • Adaptive Data Collection is OK!
    • Censored Data is OK!

    • Applies to Gaussian Processes, Brownian Motion
      • Hence, for the limiting distribution of max-likelihood estimates with i.i.d data
      • Or, Cox regression under a proportional hazards assumption

39 of 56

Assumptions

  •  

40 of 56

A workflow issue?

  •  

41 of 56

Solution:

  • Tune our rejection threshold while we perform the validation to guarantee an overall .025 bound

  • Key Idea: Use a monotone family of rejection rules, and use only one set of Monte Carlo simulations to find the rejection rule which exactly hits the .025 bound for that set of sims

(Details in thesis)

42 of 56

Suggestion for discussion:

  • Should a ledger be developed for recording and locking in the outcomes of Monte Carlo simulations?

  • This would prevent any gaming involving re-running of simulations.

43 of 56

Flexible re-design

  • Having an accurate understanding of Type I Error means we could apply conditional error Type I Error arguments to multi-dimensional hypothesis spaces

  • So, in theory, one could change to a new design which is provably below the old design’s Type I Error profile

  • A variation of this argument lets us add new arms to the trial. Perhaps this is usable for platform trials

(Further details in thesis)

44 of 56

Open floor for questions

45 of 56

Questions for you

  • What part of this work appeals to you the most?

  • What problems would you suggest to focus on next?

  • Recommendations for proof-of-concept applications?

46 of 56

Possibility: A high-speed validation pipeline

Can we set out a large class of designs, where software validation of Type I Error removes the need for any further negotiation and mathematical review by FDA statisticians?�

- “Always approved” model classes

-Binomial Outcomes

-Asymptotic gaussian statistics

-Depending on context, a class of survival models:

-proportional hazards (for log-rank and cox models)

-exponential or gamma survival distributions

47 of 56

END OF SEPT 8th TALK. �See presenter notes on this slide for detailed notes on the seminar and followup discussion

48 of 56

Returning to the math

49 of 56

Underlying Idea: Taylor Expansion

  •  

where

50 of 56

Monte Carlo on grid points

Can get estimates with Monte Carlo

51 of 56

 

  •  

52 of 56

 

  •  

53 of 56

 

54 of 56

What about ?

Martingales + Upper Bounds on Sample Sizes

Bound on the covariance matrix of

 

 

55 of 56

What about ?

 

56 of 56

Further questions?

  • End slides