Open access peer-reviewed chapter

Efficacy Evaluation in the Era of Precision Medicine: The Scope for AI

By Dominic Magirr

Submitted: August 21st 2019Reviewed: November 11th 2019Published: September 9th 2020

DOI: 10.5772/intechopen.90455

Downloaded: 33

Abstract

Patient stratification and the use of real-world evidence in regulatory decision-making are two key areas where algorithms are having an impact on drug development. The two are linked: increased patient stratification makes it harder to recruit patients into randomized-controlled trials, increasing the pressure on drug developers to find alternative sources of evidence for showing efficacy. In addition to real-world evidence, we are also seeing the emergence of more efficient ‘master protocol trials’, where multiple targeted agents can be evaluated simultaneously. In this chapter, I will review these developments and investigate the limitations for AI in terms of demonstrating the efficacy of novel targeted agents.

Keywords

  • drug development
  • precision medicine
  • statistics

1. Introduction

The use of algorithms to find patterns and make predictions from multiple data sources—here referred to as artificial intelligence (AI)—is having an increasingly large impact on clinical drug development.

Algorithms can be applied to combined clinical and genetic data sets to stratify patient populations into subgroups, based on shared characteristics or similar prognostic profiles [1, 2]. This would appear to make sense, since the majority of new drugs approved by the US FDA in recent years have been targeted towards specific genetic aberrations [3, 4]. If we increase our search, we will find more genetic aberrations, more drug targets, and more potentially efficacious drugs. However, this approach also presents severe challenges in the clinical stages of drug development, as the size, complexity and duration of studies increases.

One way to react to increased cost and duration is to improve the operational efficiency of clinical trials. The last decade has seen the emergence of ‘master protocol trials’, which allow several substudies to be conducted simultaneously, reducing the rate of screen failures [5]. In addition, there is increasing enthusiasm for augmenting (possibly even replacing) randomized-controlled trials (RCTs) with external and real-world data, where it is claimed that further use of algorithms can protect us from the biases that this approach would otherwise impose [1].

The purpose of this article is three-fold. Firstly, to explain how precision medicine presents challenges to traditional drug development, quantifying the effect of disease stratification on trial recruitment. Secondly, to describe how master protocol studies have emerged in response to these challenges. Finally, to explore whether it is possible for single-arm studies with ‘synthetic control arms’ to provide the same standard of evidence as a randomized controlled trial, thus reducing drug development timelines.

2. Disease stratification

Consider a patient population that can be stratified according to the value of a diagnostic test. The ‘target’ population consists of patients who test positive. The ‘non-target’ population consists of patients who do not test positive. Suppose that a new treatment is expected to be more effective in the target population than in the non-target population. Let θ+and let θdenote the treatment effect sizes in target and non-target populations, and γdenote the prevalence of the target group. Three things that we would like to demonstrate are:

  1. Treatment benefit in the full population, γθ++1γθ>0.

  2. Treatment benefit in the target population, θ+>0.

  3. Greater benefit in the target population than in the non-target population, θ+>θ.

Which of these is easiest to demonstrate, and which most difficult? To answer this, we compare the standardised statistics, Z, that we would use to test the corresponding null hypotheses. For most commonly-used clinical-trial endpoints, the test statistic ends up looking like

ZNθI1,E1

where θis the treatment effect size and Iis the statistical information, which is typically proportional to the sample size [6, 7]. The power of a test is the probability that Z>k, for a threshold k, where kis chosen to ensure a given false-positive rate. The larger the expected value of Z, the higher the power. Therefore two trials (‘A’ and ‘B’) will have the same power if θAIA=θBIB, or, assuming that information is proportional to sample size, if

θANA=θBNB.E2

We can use (2) to assess the relative difficulty of our three goals, firstly for the full population versus the interaction (1. versus 3.), and then for the full population versus the target population (1. versus 2.).

2.1 Full population versus interaction

It is shown in the appendix that a test of the interaction null hypothesis, θ+=θ, with total sample size Nint, will have the same power as the test for the full population null hypothesis, γθ++1γθ=0, with sample size N, provided that

θ+θγ1γNint=γθ++1γθN.E3

For example, when θ/θ+=0.5, for a prevalence of 50%, the ratio of sample sizes is Nint/N=9. For a prevalence of 5%, Nint/N23. This shows how difficult it is to provide compelling evidence for treatment-biomarker interactions, and why drug development is still focussed on demonstrating average treatment effects. It is also explains why post-hoc data-driven subgroup identification following a clinical trial is often a bad idea. See Gelman [8] for further discussion.

2.2 Full population versus target population

A test for the full population null hypothesis with sample size Nwill have the same power as a test for the target population null hypothesis, θ+=0, with sample size NT, provided that

γθ++1γθN=θ+NT,E4

or, equivalently, if

NTN=γ+1γθθ+2.E5

In (5), we have expressed the relative sample size, NT/N, as a function of the relative efficacy, θ/θ+[9]. This relationship is drawn in solid lines in Figure 1 for two potential prevalences (50% and 5%) when θ/θ+is between 0.5 and 1. For a prevalence of 50%, the targeted strategy requires up to 40% fewer patients than the non-targeted strategy. For a prevalence of 5%, a 70% reduction is possible. Note, however, that this is the relative number of patients enrolled. What about the number of patients screened? In the full population the minimum number screened is N, whereas in the targeted population it is Ntdivided by γ. The ratio, Nt/γ/N, is drawn in dashed lines in Figure 1. For the 50% prevalence case, there is a maximum 2-fold increase in the number screened for the targeted compared to the non-targeted trial. But for the 5% prevalence case, there is somewhere between a six-fold and a twenty-fold increase.

Figure 1.

Relative sample size when testing efficacy in the target population compared to the full population (NT/N), shown in solid lines. The dashed lines show the relative number of patients screened (NT/γ/N).

2.3 Situations where θθ+

The conclusion from Figure 1 is that population stratification is only likely to be useful if there exists a potential treatment where the treatment effect is considerably higher (e.g. at least two-fold) in the target subgroup than in the rest of the population. Marginal increases in efficacy are not enough in practice. The targeted approach would require a prohibitively large number of patients to be screened, compared to a trial in the full population which would have the same statistical power. Marginal increases are also difficult to establish empirically, as shown in Section 2.1. It follows that successful implementation of precision-medicine drug development is restricted to situations where there is strong biological and pre-clinical evidence for expecting θθ+. Such cases certainly do exist, and the targeted trial is the only sensible approach here. Nevertheless, one still needs to screen a very high number of patients. This is expensive for the sponsor. It is also disheartening for patients who do not meet the eligibility criteria.

3. Master protocol trials

The high screen failure rate of precision-medicine trials can be mitigated to some extent by merging multiple sub-studies into a single ‘master protocol’. The last decade has seen the emergence of the labels ‘basket’ and ‘umbrella’ to describe these complex studies. As a rule of thumb, a basket tends to refer to studies involving the same drug in multiple diseases, whereas umbrella is used when multiple experimental treatments are studied in the same disease. However, as reported by Janiaud and colleagues [10], these terms have not been applied consistently. Their systematic review of master protocol trials in oncology found 30 ‘basket’ trials and 27 ‘umbrella’ trials in a time period of 2006–2018, but with most studies starting after 2015. They explain that some basket trials are mistakenly labeled as umbrella trials, and vice-versa, but there are also trials that contain elements of both and thus become difficult to describe using current language.

Stallard and colleagues [11] propose a refined classification which replaces ambiguous labels with a more precise visual description, as shown in Figure 2. In each of the six designs, a small square is representative of a cohort of patients. On the left hand side are the basket-type designs, where there is only one new treatment (T) targeting a particular mutation (M), but this mutation occurs across diseases (D1,D2,). In the middle are the umbrella-type designs, where there are multiple treatments (T1,T2,) targeting particular mutations (M1,M2,), all within the same overall disease (D). The designs on the right hand side combine the features of the basket-type and umbrella-type designs. They allow for multiple disease types within each of the separate treatment-mutation combinations. Note, however, that it is always the mutation that is driving the choice of treatment, rather than the disease type. In all of the designs, for each T-M-Dsub-study, it is possible to use a single-arm design (Figure 2a), or compare with a concurrent control arm (Figure 2b).

Figure 2.

A classification of master protocol designs by Stallard and colleagues [11]. Each square represents a cohort of patients. (a) Single-arm cohorts. (b) Concurrent control arms.

3.1 Example 1: Vemurafenib in cancers (not melanoma) with BRAF V600 mutations

Hyman and colleagues [12] report the results of a basket-type study with the same structure as the left-hand-side of Figure 2a. The treatment (T) was Vemurafenib, the mutation (M) was BRAF V600. There were several cohorts corresponding to different disease types (D):

  • Colorectal cancer (CRC)

  • Bile duct cancer

  • Anaplastic thyroid cancer (ATC)

  • Non-small cell lung cancer (NSCLC)

  • Erdheim-Chester disease/Langerhans cell histiocytosis (ECD/LCH)

A Simon’s two-stage design [13] was used for each cohort independently to allow for early futility stopping. Consequently, the cohort sizes ranged from 5 to 27. Hobbs and colleagues [14], did a re-analysis of the data, and their findings are reproduced in Figure 3. Looking at the response rate across cohorts, it appears that there is more activity in NSCLC and ECD/LCH than in CRC. However, one can also see a clear inverse relationship between response rate and number of prior therapies, which muddies the water. This example highlights how difficult it can be to interpret uncontrolled studies.

Figure 3.

Results from a basket-type study of Vemuarafenib [12, 13].

3.2 Example 2: FOCUS4

An example of a master protocol trial that does include concurrent control arms is FOCUS4 [15], currently being run by the Medical Research Council Clinical Trials Unit in London. It has an umbrella-type design like Figure 2b. The disease setting (D) is advanced colorectal cancer. Mutations (M) include:

  • BRAF mutations

  • MSI deficient

  • PIK3CA mutations

  • Wild type

A centralised molecular analysis is performed on each patients tumor. Based on the results, patients are offered entry into an appropriate substudy, where they are randomized to receive either an experimental treatment (T) targeted to their mutation, or a control treatment.

The substudies will be analysed independently, as if they were separate trials. The big advantage over independent studies is the increased efficiency from the centralised molecular analysis, ensuring fewer screen failures. Complications may arise when patients are eligible for more than one substudy, and this has to be planned for in the protocol. Note also the inclusion of the Wild-type cohort in FOCUS4. This maximizes the proportion of patients who undergo screening who are given an option to go on a trial.

4. External control arms

Precision medicine is increasing the pressure on drug developers to find innovative ways to demonstrate efficacy without requiring ever larger and lengthier clinical trials. We have seen how operational efficiencies can be found in master protocol trials. A related development is the use of ‘big data’—the bringing together of historical RCT data, electronic health records, advanced statistical modeling, and machine-learning—to produce a historical benchmark, or even a so-called ‘synthetic control arm’, that might allow a single-arm study to take the place of an RCT as a basis for seeking drug approval.

For this approach to be successful, the key use-case in oncology is a comparison of overall survival (OS). It is typical for inference to focus on the (log) hazard ratio,

θlogλEtλCt,E6

where it is assumed that the hazard of death on the experimental arm, λEt, is proportional to the hazard of death on the control arm, λCt, for all timepoints t. Another way to describe λtis that it is your risk of dying on day tgiven that you were alive at midnight. More stringent than proportional hazards is an assumption of constant hazards, λjt=λjfor all t(j=E,C). Although an over-simplification, this model is often not a bad approximation to reality, and we will use it to compare operating characteristics for a two-arm RCT versus a single-arm trial with an external control arm.

4.1 Distribution of treatment effect estimators

The constant-hazards assumption allows us to express the log hazard ratio as the difference between the log-transformed median survival times,

logλEλC=logmClogmE.E7

For a two-arm study with equal randomisation and Devents, the estimate of the log hazard ratio has the following (approximate) distribution:

θ̂=logm̂Clogm̂ENθ4/D.E8

If we were to run a single-arm study instead, but keep the overall sample size the same, i.e. put all patients who would have received the control treatment onto the experimental arm, we could use the test statistic

θ̂=logmClogm̂ENθ+logmClogmC1/DE9

where mCis our best pre-trial estimate for the median OS on the control arm.

4.2 Bias-variance trade-off

We can compare the precision of the two estimates in terms of their mean-squared-errors,

mseθ̂=varθ̂+biasθ̂2=4/D+0E10

and

mseθ̂=varθ̂+biasθ̂2=1/D+logmClogmC2.E11

For low values of D, variance will be a bigger problem than bias. In this case, mseθ̂<mseθ̂. However, as soon as

D>3logmClogmC2E12

the bias will dominate, and the estimate from the two-arm trial will be more precise.

4.3 NSCLC example

What is a typical value for logmClogmC? This depends on the context. The FDA have published data from 14 large randomized control trials [16] in advanced non-small-cell-lung cancer (NSCLC) conducted between 2003 and 2015. The median survival on the control arm across the studies is shown in Figure 4. Three of the studies were targeted towards patients with a particular biomarker. It is immediately obvious that these three data points are different from the rest, and this highlights the dangerous territory we are in. Nevertheless, if we focus on the 11 studies that did not use a targeted approach, the median overall survival ranged from 7 to 13 months. Taking an average value, a sensible choice for logmCis log9.5. We could also think about the ‘true’ logmCfor the current study belonging to the same distribution as the 11 other studies, which we might approximate with a normal distribution

Figure 4.

Between-trial variability in median overall survival time from 14 phase 3 studies in advanced non-small-cell lung cancer trials submitted to the FDA [16].

logmCNlogmC=log9.5σmC2=0.03E13

The expected value of logmClogmCaccording to (13) is 2/πσmC0.14. Plugging this into (12), the two-arm trial would be more precise than the single-arm trial when D>153.

4.4 Reducing the sample size

What if instead of moving patients from the control arm to the experimental arm and keeping total sample size the same, we run a single-arm study with half the number of patients, i.e. we keep the same sample size on the experimental arm and replace the control arm with an historical benchmark? In this case, the mean-squared-error of the estimate from the two-arm trial will be lower than the single-arm equivalent as soon as

D>2logmClogmC2,E14

where Dis the number of events in the two-arm trial. For our lung cancer example, this would mean as soon as D>102.

4.5 More advanced methods

In the previous example we were using the average value from 11 previous studies as a rather crude estimate of logmC. Is it possible to improve the precision using ‘big data’—bringing together historical RCT data, electronic health records, advanced statistical modeling, and machine-learning?

We can look to a recent study by Carrigan and colleagues [17]. The group had access to individual patient data from 9 RCTs in advanced NSCLC conducted between 2011 and 2018, as well as electronic health records (EHR) from almost 50,000 patients. They used advanced regression and stratification techniques to estimate treatment effect sizes, and their results are reproduced on the left hand side of Figure 5. There is a high correlation (0.86) between the hazard ratio from the RCTs and the hazard ratio that would have been observed had the control arm been replaced with electronic health record data. On the right hand side, the data points have been transformed into an estimate of the bias logmClogmC, assuming constant hazards. The mean value is 0.1 and according to (12) this means that a two-arm trial would be more precise than a single-arm trial of the same total sample size whenever D>300. Similarly, using (14), a two-arm trial will be more precise than a single-arm trial with half the sample size when D>200.

Figure 5.

Correlation between RCT-derived and EHR-derived hazard ratios from nine studies in advanced non-small-cell lung cancer [17]. On the right-hand-side, the results have been converted into an approximation of the bias when estimating the median survival time on the control arm using EHR data.

To put these findings in some context, for a study with one-sided type-1 error of α=0.025, 300 events would give 90% power when HR=0.69. Likewise, 100 events would give 90% power when HR=0.52.

5. Conclusions

Advances in pattern-recognition and prediction algorithms have the potential to improve health outcomes, as well as making the drug development process more efficient. Nevertheless, it is important to have a strong grasp of some limiting factors, to avoiding spending time on futile endeavors.

The stratification of patient populations into ever finer subgroups is only likely to prove useful when there exist potential treatments with very large differential treatment effects. Marginal is not enough—it needs to be 100% more efficacious in the target subgroup than in the non-target subgroup. Otherwise, a clinical trial in the full population would have the same statistical power with far fewer patients screened. This means that we need strong biological rationale and robust pre-clinical evidence. In addition, it is essential that the diagnostic test has high sensitivity and specificity. Otherwise, a large treatment effect in the true biomarker-positive population would become diluted in the observed biomarker-positive population.

In cases where there is a strong rationale for a targeted approach, recruitment will be challenging. Master protocol trials can be an excellent option. They are an efficient way to test novel agents, and they increase the chance that a patient entering screening will be able to join a clinical trial.

Improvements in the quality of electronic health records, as well as better algorithms to interrogate this data, are a positive development that can enhance our understanding of health outcomes, and help enormously with clinical trial design and interpretation. Nevertheless, we should not forget the fundamental benefits of concurrent control [18], and should remain realistic about the ability of synthetic control arms to replace the real thing. We have seen that under favorable circumstances (highly prevalent disease, patient-level data from numerous high-quality large RCTs, tens of thousands of electronic health records, well-defined and accurately-measured primary endpoint, careful analysis), a single-arm study can provide similar precision to a two-arm randomized comparison with sample size in the low hundreds [17]. It is plausible, therefore, that for a new drug in this space with a very large treatment effect, a single-arm study may provide convincing evidence of efficacy. But one should expect this to be the exception, not the norm.

Conflict of interest

Dominic Magirr is an employee of Novartis Pharma AG.

Appendix

Based on the test statistics (1) for the target and non-target populations,

Z+Nθ+γIint1

and

ZNθ1γIint1,

we can define an interaction test statistic

Zint1γZ+γZNθ+θγ1γIint1.

By (2), this test will have the same power as the full population test with sample size Nif

θ+θγ1γNint=γθ++1γθN.

Abbreviations

AIartificial intelligence
FDAFood & Drug Administration
RCTrandomized controlled trial
NSCLCnon-small-cell lung cancer
OSoverall survival
EHRelectronic health record

© 2020 The Author(s). Licensee IntechOpen. This chapter is distributed under the terms of the Creative Commons Attribution-NonCommercial 4.0 License, which permits use, distribution and reproduction for non-commercial purposes, provided the original is properly cited.

How to cite and reference

Link to this chapter Copy to clipboard

Cite this chapter Copy to clipboard

Dominic Magirr (September 9th 2020). Efficacy Evaluation in the Era of Precision Medicine: The Scope for AI, Artificial Intelligence in Oncology Drug Discovery and Development, John W. Cassidy and Belle Taylor, IntechOpen, DOI: 10.5772/intechopen.90455. Available from:

chapter statistics

33total chapter downloads

More statistics for editors and authors

Login to your personal dashboard for more detailed statistics on your publications.

Access personal reporting

Related Content

This Book

Artificial Intelligence in Oncology Drug Discovery and Development

Edited by John Cassidy

Next chapter

AI Enabled Precision Medicine: Patient Stratification, Drug Repurposing and Combination Therapies

By Steve Gardner, Sayoni Das and Krystyna Taylor

Related Book

First chapter

Nonparametric Modeling and Model-Based Control of the Insulin-Glucose System

By Mihalis G. Markakis, Georgios D. Mitsis, George P. Papavassilopoulos and Vasilis Z. Marmarelis

We are IntechOpen, the world's leading publisher of Open Access books. Built by scientists, for scientists. Our readership spans scientists, professors, researchers, librarians, and students, as well as business professionals. We share our knowledge and peer-reveiwed research papers with libraries, scientific and engineering societies, and also work with corporate R&D departments and government entities.

More About Us