Chapter 9 Causal inference for precision medicine
Felix qui potuit rerum cognoscere causas.
Virgil (Georgics, 29 BC)
Throughout this manuscript, we first described the complexity of cancer mechanisms, through the diversity of genetic alterations or non-linear signaling pathways. This complexity naturally led to the choice of systemic modeling approaches and in particular mechanistic models whose explicit nature facilitates the study of the effects of new molecular perturbations such as treatments. The simple study of the response to BRAF inhibitors has thus required the consideration of many other genes and pathways.
This final chapter proposes to take the complexity a step further by considering different treatments. The diversity of patients' molecular profiles suggests that the best treatment is not necessarily the same for all patients: this is what is known as precision medicine. This is already a clinical reality in oncology that could be reinforced in the future by the emergence of new computational models of cancer, whether mechanistic or not. How then can we assess the relevance of these models in their ability to guide patient treatment?
Scientific content
This chapter presents an extension of the causal inference framework to quantify the value of precision medicine strategies. This work is currently under revision and is available as a preprint in Béal and Latouche (2020). All code is available in the dedicated GitHub repository
9.1 Precision medicine in oncology
It is first important to understand what is meant by the concept of precision medicine in the treatment of cancer patients in order to place subsequent questions in a plausible clinical framework.
9.1.1 An illustration with patient-derived xenografts
Precision medicine stems from the diversity of treatment responses observed in different tumors. It has already been observed in previous chapters about BRAF inhibition that different cell lines respond differently to a particular treatment. A broader analysis of pre-clinical data shows that the same is true for the vast majority of treatments. It would be possible to illustrate this using the same data from cell lines extended to other drugs. However, because of the more directly clinical impact of the issues discussed in this chapter, the analyses presented below will focus on another type of data that is closer to patient data: patient-derived xenografts (PDX).
A PDX is a tumor tissue that has been removed from a patient and implanted into immunodeficient mice (Hidalgo et al. 2014). Unlike cell lines, which are in vitro models, PDXs are in vivo models that allow cancer cells to evolve in a more realistic microenvironment. In the same way as for cell lines, PDX can be used for drug screening. The data used in this chapter come from a study by Gao et al. (2015) which contains several hundred tumors and more than fifty drugs. Not all drugs have been tested for all tumors; details of the drugs and types of cancer tested are available in the appendix A.2. This dataset was generated following the "one animal per model per treatment" approach (\(1 \times 1 \times 1\)), the principles of which are summarized in Figure 9.1A. It should be noted that different drug response metrics are computed in the source data, two of which will be used in the analyses. The first one is continuous and called Best Average Response in the data, it is based on the variation of the tumor volume after treatment, the lower values (and especially negative) corresponding to better responses. The second one is originally categorical and based on a modified Response Evaluation Criteria In Solid Tumors (RECIST) criteria. It was binarized for this study so that the responders have a score of \(1\) and non-responders \(0\). The details of the definition and distribution of these metrics are given in appendix A.2.
In order to illustrate the diversity of response to treatment, the database is momentarily restricted to the 4 most widely tested drugs and the 180 tumors (or PDX models) that were evaluated for all four drugs. The four chosen drugs target different pathways: binimetinib (MAPK inhibitor), BKM120 (PIK inhibitor), HDM201 (MDM2 inhibitor) and LEE011 (CDK inhibitor). In Figure 9.2A the 4 treatments show a high variability of response, with a slight advantage for BKM120 and binimetinib on average over all tumors. However, each of the treatments was found to be the most effective of the 4 for a significant proportion of tumors (Figure 9.2B), with binimetinib and BKM being the best treatment for one-third of tumors each and LEE011/HDM201 sharing the remaining one-third of tumors. It thus appears that in view of the molecular diversity of tumors and the increasing number of treatments available, it does not seem advisable, according to these preclinical data, to treat all tumors with the same gold-standard treatment. Furthermore, the tissue of origin of the tumors in this example does not appear to be the main determinant of tumor preference for certain treatments.
9.1.2 Clinical trials and treatment algorithms
These remarks can be extended to patients. Thus, precision medicine (PM) consists in assigning the most appropriate treatment to each patient according to his or her characteristics, usually genomic alterations for cancer patients (Friedman et al. 2015; De Gramont et al. 2015). At the individual level, targeted treatments have provided relevant solutions for patients with specific mutations (Abou-Jawde et al. 2003). Putting together these various treatments, some precision medicine strategies can be defined. Based on the genomic profile of the patient, the treatment most likely to be successful is chosen. If the information available is reliable, precision medicine can thus be reduced to a treatment algorithm that takes as input the molecular characteristics of the patient's tumor and outputs a recommendation of treatment. An example of such a treatment algorithm from the SHIVA clinical trial by Le Tourneau et al. (2015) is shown in Figure 9.3 where different treatments are associated with different alterations. In this case, the treatment algorithm can be considered as an aggregation of the medical knowledge accumulated on the individual biomarkers.
9.1.3 Computational models to assign cancer treatments
The treatment algorithm example in Figure 9.3 could, however, be more complex. Indeed, previous chapters have stressed, for example, that being mutated for the BRAF gene is not the only predictor of response to an inhibitor of BRAF (here Vemurafenib). The same is true for most treatments that could benefit from more global and systemic analyses, taking into account more variables and their interactions. This complexity would require the use of computational methods.
It is on this point that this chapter links to the previous ones. Some of the cancer models studied throughout this thesis, or their future developments, could be interpreted as treatment algorithms. Indeed, a model capable of predicting the response to a single treatment does not necessarily allow the inference of precision medicine strategies. On the other hand, a model capable of predicting a patient's response to different treatments is also capable of indicating which one is the best. Such models would then move from systems biology to systems therapeutics (Hansen and Iyengar 2013), taking patients' genomic features as inputs and outputting a treatment recommendation. In theory, mechanistic models seem to be suitable for this purpose since their explicit representation of genes and proteins makes it possible to simulate the effect of different therapeutic interventions. However, the feasibility of designing and calibrating such a model has yet to be demonstrated. Other types of models are being studied that could achieve these goals. For example, some recent approaches propose the use of deep learning to provide a computational tool for predicting the growth of cells (Ma et al. 2018), or even the sensitivity of cell lines to different treatments (Manica et al. 2019).
In short, if no computational model is sufficiently developed to date to replace the clinician, the emergence of this type of tool is likely in the medium term. This raises the question of how to assess the clinical value of the precision medicine strategies (and corresponding treatment algorithms) derived from these models. For the sake of generality, this question will be addressed more broadly in the following without reference to models as a possible source of the treatment algorithm: how to evaluate the clinical impact of a precision medicine strategy and the treatment algorithm? The methods presented will indeed be the same, whether the algorithm evaluated comes from a model or from the knowledge of clinicians as in Figure 9.3. In the spirit of this thesis, the question nevertheless finds its origin in the first hypothesis related to models.
9.2 Emulating clinical trials to evaluate precision medicine algorithms
9.2.1 Objectives and applications
The question then arises of how to quantify the clinical benefit provided by these treatment algorithms. Some precision medicine clinical trials have been proposed, demonstrating both the feasibility of collecting information about mutations (Le Tourneau et al. 2015) or RNA (Rodon et al. 2019) in real-time and the clinical benefit that can be expected from these approaches for some patients (Coyne, Takebe, and Chen 2017). However, the increasing abundance of genomic data and biological knowledge make it progressively easier to establish new algorithms for precision medicine, either directly based on physician knowledge or provided by computational models (Hansen and Iyengar 2013). For practical reasons it is not possible to propose a real clinical trial for each new precision medicine algorithm or for any variants, comparing standard of care with new algorithm-based therapeutic strategies.
Therefore, this work provides a method to assess the clinical impact of proposed PM treatment algorithm based on already generated data, emulating precision medicine clinical trials and analyzing them in the causal inference framework (Hernán and Robins 2016). First we will define the causal estimates of the precision medicine effects (later referred to as causal estimates) we want to assess, and the corresponding ideal clinical trials one would like to perform. Next, we will define the notations and the causal framework we use to infer the causal effects from observational data with multiple versions of treatment, based on the previous work by VanderWeele and Hernan (2013). The main principles of the potential outcome framework having been introduced in the previous chapter, an extension to the case of precision medicine will be described, focusing on the multiplicity of treatment versions, i.e., targeted drugs. Then we will apply these methods to simulated data in order to investigate the different biases of the candidate methods. An example scenario will be presented and a RShiny interactive application has been developed to further explore other user-defined settings. Finally, the analysis of data from patient-derived xenografts (PDX) makes it possible both to apply the methods to pre-clinical situation and to have data approximating the counterfactual responses, thus enabling further validation of the proposed estimation methods.
9.2.2 Target trials for precision medicine: definition of causal estimates
We first specify the precision medicine effects that are to be estimated. These effects will finally be estimated based on observational data through the causal framework and target trial emulation (Hernán and Robins 2016). In this context the notion of target trial refers to the real clinical trial whose estimates are sought to be reproduced through causal inference. Thus, if we think in terms of clinical trials, we are not trying to prove or quantify the superiority of one treatment over another but rather to evaluate the clinical utility of a precision medicine strategy assigning treatments based on genomic features of patients. This is therefore closer to the well-studied biomarker-based designs for clinical trials (Freidlin, McShane, and Korn 2010). In a way, it is a matter of extending these unidimensional biomarker-based designs to multidimensional strategies that allow a choice between quite a number of different treatments. The potentially large number of treatments thus prompts us to draw more inspiration from scalable biomarker-strategy designs than biomarker-stratified designs (Freidlin, McShane, and Korn 2010). We can draw a methodological parallel with some trials like the Tumor Chemosensitivity Assay Ovarian Cancer study in which a biochemical assay guides the choice of preferred chemotherapy for patients in a panel of twelve different treatments (Cree et al. 2007). More recently, some clinical trials have been proposed that include precision medicine strategies, particularly in oncology (Von Hoff et al. 2010; Le Tourneau et al. 2015; Flaherty et al. 2020).
On the basis of these clinical examples, we propose three different target trials and their corresponding causal estimates, the clinical relevance of which may vary according to medical contexts. Each target trial contains a precision-medicine directed arm in which patients are treated in accordance with the precision medicine algorithm recommendations but they are differentiated from each other by alternative control arms (Figure 9.4). Causal effects will be estimated solely on patients eligible for the assignment of a personalized treatment, i.e., those for whom the treatment algorithm is able to recommend a drug.
9.2.2.1 First causal effect (\(\text{CE}_1\)): comparison with a single standard
The first possible target trial is to compare the precision medicine arm with a control arm in which all patients have been treated with the same single treatment. This could classically be the current standard of care applied to all patients (e.g., chemotherapy cancer treatment).
9.2.2.2 Second causal effect (\(\text{CE}_2\)): comparison with physician's assignment of drugs
Then, in order to propose a more comprehensive clinical assessment, we propose a second causal effect, comparing the PM arm with the current clinical practice, i.e., the assignment of the same targeted treatments by physicians in the absence of the algorithm. This implicitly means comparing two PM strategies: the one derived from the algorithm and the one that corresponds to current physician's knowledge. Unlike the former, the latter may not be perfectly deterministic depending on the heterogeneity of medical knowledge or practices. This way of defining \(\text{CE}_2\) by focusing on the doctor's assignment of the same treatments stems from our question of interest: to quantify the relevance of the algorithm itself. Another possibility would have been to compare the precision medicine arm with the doctor's treatments, allowing him to use treatments other than those of the PM arm, such as the gold-standard one described in \(\text{CE}_1\). But the differences between the arms could then be biased by the use of treatments with different overall efficacy, changing the focus of the question. We will therefore stick to the first definition, which is more focused on the relevance of the algorithm.
9.2.2.3 Third causal effect (\(\text{CE}_3\)): comparison with random assignment of drugs
Finally, we define the \(\text{CE}_3\) effect comparing the PM arm with a control arm using exactly the same pool of treatments assigned randomly. In this case, we measure the ability of the PM algorithm to assign treatments effectively based on genomic features of patients. This comparison has already been considered in the context of biomarker-based clinical trials (Sargent et al. 2005). Although this comparison with random assignment is methodologically relevant, it may not make sense from a clinical point of view if the common clinical practice already contains strong indications (or contraindications) for some patient-treatment associations.
9.3 Causal inference methods and precision medicine
9.3.1 A treatment with multiple versions
The statement of the potential outcomes framework implicitly implies the uniqueness of the versions of the treatment (Rubin 1980) or at least the treatment variation irrelevance (VanderWeele 2009). In the precision medicine case, the multiplicity of treatment versions is inherent: a given treatment status may encompass several drugs since a patient may be associated with several molecular agents based on his or her genomic characteristics. \(A\) can be seen as a compound treatment (Hernán and VanderWeele 2011) or a treatment with multiple versions (VanderWeele and Hernan 2013).
Therefore, we define a variable \(K_j\) denoting the version of treatment administered to individual \(j\). If \(A_j=a\) is the arm to which the patient is assigned, \(K^a_j\) is the molecule received, the version of treatment \(A=a\) (e.g., a specific anti-cancer drug) and \(K^a_j \in \mathcal{K}^a\), the set of versions of treatment \(A=a\). In our precision medicine problem, \(A=0\) will denote control patients and \(A=1\) the patients treated with an anti-cancer drug of the precision medicine pool. \(\mathcal{K}^1=\{k^1_1, ..., k^1_P\}\) is the set of \(P\) possible targeted treatments for \(A=1\) patients. For the sake of simplicity we will assume that there is only one treatment version for \(A=0\) controls, \(\mathcal{K}^0=\{k^0\}\). We also need to define other counterfactual variables like \(K^a_j(a)\), the counterfactual version of treatment \(A=a\) if the subject had been given the treatment level \(a\). Thus, we finally write the counterfactual outcome as \(Y_j(a,k^a)\) for individual \(j\) when treatment \(A\) has been set to \(a\), using \(k^a\) as the version of treatment \(a\), with \(k^a \in \mathcal{K}^a\). Causal relations between variables \(C\), \(A\), \(K\) and \(Y\) are depicted in the causal diagram in Figure 9.5. It should be noted that \(A\) has no direct influence on \(Y\), its only effect is entirely mediated by \(K\), which is the real treatment in the pharmacological sense.
In this context, we can also define the assignment of a version of treatment for patients eligible for precision medicine algorithm. It is important to note that not all patients are necessarily eligible for the precision medicine strategy. Indeed, the treatment assignment algorithm relies on targetable alterations to establish its recommendations. In the absence of these, no recommendation can be offered to the patient. We denote \(\mathcal{C}^{PM}\) the set of eligible patient profiles and consequently define the drug assignment algorithm as the function \(r\) which associates to each \(C\) a precision medicine treatment version \(K\) such as:
\[\forall j \in [[ 1, N ]], \: \textrm{if} \: C_j \in \mathcal{C}^{PM}, r(C_j) \in \mathcal{K}^{1}\]
9.3.2 Causal inference with multiple versions
Consequently, the multiplicity of versions prevents direct application of the framework as described in section 8.2.2. The theoretical framework has however been extended to causal inference under multiple versions of treatment and some identifiability conditions and properties have been studied, especially in the seminal article by VanderWeele and Hernan (2013). One of the first required adaptation to identify some causal effects is to distinguish between confounders \(C\) and \(W\) (Figure 9.5). \(W\) indicates a collection of covariates that may be causes of treatment \(A\) or version of treatment \(K\) but are not direct causes of \(Y\). These covariates are of special interest for causal effects identification under multiple versions of treatment. \(C\) indicates all other covariates. In our precision medicine settings, the genomic features of patients may define the eligibility to precision medicine and therefore affect \(A\). They may also be used to define the version of treatment \(K\). And finally they can influence the response to treatment \(Y\). Thus, the genomic features of patients are a typical example of type \(C\) confounders. All causal relationships are summarized in Figure 9.5. Please note that no \(W\) variable is present in the applications provided later because all the covariates considered in this situation were likely to influence \(A\), \(K\) and \(Y\) and therefore belonged rather to the covariates of type \(C\). However all subsequent formulas and definitions have been derived taking into account \(W\).
We summarize here some general observations from VanderWeele and Hernan (2013) regarding the extension of the framework to multiple versions before discussing specific estimates of interest of our precision medicine settings in the next section. These two sections will be based exclusively on the method called standardization or parametric g-formula described in section 8.2.2.1. The adaptation of other methods to precision medicine will be discussed more briefly in section 9.3.4. First of all, the identifiability conditions have to be adapted. The consistency assumption for instance is extended to \(K\):
\[\textrm{if} \: A_j=a, \textrm{then} \: K_j^a(a)=K_j^a\].
Then, the conditional exchangeability or no-unmeasured confounding assumptions, may be stated in two different ways, either without or with versions of treatment:
\[\begin{equation} Y(a) \perp \!\!\! \perp A | (C,W) \tag{9.1} \end{equation}\] \[\begin{equation} Y(a, k^a) \perp \!\!\! \perp \{A, K\} | C \tag{9.2} \end{equation}\]If equation (9.1) holds, we can derive a new version of the standardised estimator with multiple versions of treatment (VanderWeele and Hernan 2013):
\[\begin{equation} E[Y(a)] = E[Y(a, K^a(a))] = \sum_{c,w} E[Y | A=a, C=c, W=w] \times P[c,w] \tag{9.3} \end{equation}\]Specifically, it should be noted that we need to add \(W\) in the set of covariates that must be taken into account in standardization, and we need positivity to hold for \(C\) and \(W\), i.e., \(0<P[A=a|C=c,W=w]<1\). Detailed proof of equation (9.3) is provided in appendix C.2.1. Equation (9.3) paves the way to overall treatment effect assessment since \(E[Y(1, K^1(1))]-E[Y(0, K^0(0))]\) would estimate the effect of treatment \(A=1\) compared to \(A=0\) with current versions of treatment.
Conversely, estimating a treatment effect for a given unique version of treatment \(E[Y(a,k^a)]\) would require to check the exchangeability with regard to versions \(K\) and therefore to hold equation (9.2) true (VanderWeele and Hernan 2013):
\[\begin{equation} E[Y(a, k^a)] = \sum_{c} E[Y | A=a,K^a=k^a, C=c] \times P[c] \tag{9.4} \end{equation}\]Similarly, we can define \(G^a\) a random variable for versions of treatment with conditional distribution \(P[G^a=k^a| C=c]=g^{k^a,c}\) and assuming the equation (9.2) to be true we can derive the following formula and its formal proof in appendix C.2.2:
\[\begin{equation} E[Y(a, G^a)] = \sum_{c,k^a} E[Y | A=a,K^a=k^a, C=c] \times g^{k^a,c} \times P[c] \tag{9.5} \end{equation}\]In this case, to allow estimation of the right-hand side of the equation, positivity will be defined as \(0<P[A=a, K^a=k^a|C]<1\).
9.3.3 Application to precision medicine
In the context of the potential outcomes framework extended to treatments with multiple versions, it is therefore possible to apply equations (9.3) and (9.5) in order to define and estimate the precision medicine causal effects previously described in section 9.2.2.
\(A=0\) corresponds to control patients with \(\mathcal{K}^0=\{k^0\}\) and \(A=1\) to patients treated with a targeted treatment. It is important to notice that from this point on we systematically restrict ourselves to patients eligible for the precision medicine algorithm, i.e., to individuals \(j\) such as \(C_j \in \mathcal{C}^{PM}\).
9.3.3.1 \(\text{CE}_1\) estimation
\(\text{CE}_1\) is a comparison between the precision medicine arm and a single version control arm:
\[\begin{equation} \text{CE}_1 = E[Y(1, r(C)] - E[Y(0, k^0)] \tag{9.6} \end{equation}\]In details, \(E[Y(1, r(C)]\) can be derived from equation (9.5) in the case where \(g^{k^a,~c}=1\) if \(k^a=r(c)\) and \(g^{k^a,~c}=0\) otherwise:
\[E[Y(1, r(C)] = \sum_{c} E[Y | A=1,K^1=r(c), C=c] \times P[c]\]
Then, \(E[Y(0, k^0)]\) and \(E[Y(1, k^1_{ref})]\) can be derived from equation (9.4):
\[ E[Y(0, k^0)] = \sum_{c} E[Y | A=0, C=c] \times P[c] \]
Alternatively, if one wants to use as control only one of the treatments used in the PM arm the previous estimate could be replaced by the following one: \[E[Y(1, k^1_{ref})] = \sum_{c} E[Y | A=1,K^1=k^1_{ref}, C=c] \times P[c]\]
It should be noted that \(\text{CE}_1\), like \(\text{CE}_2\) and \(\text{CE}_3\) presented later, depends on the PM algorithm of interest \(r\). \(\text{CE}_i\) could therefore also be written \(\text{CE}_1(r)\).
9.3.3.2 \(\text{CE}_2\) estimation
Then, \(\text{CE}_2\) is written using \(K^1(1)\) the PM targeted treatment that would have been assigned to the patient by the physician if the patient had been allocated in arm \(A=1\) with PM targeted treatments:
\[\begin{equation} \text{CE}_2 = E[Y(1, r(C)] - E[Y(1, K^1(1))] \tag{9.7} \end{equation}\]\(E[Y(1, K^1(1))]\) is derived from equation (9.3):
\[E[Y(1, K^1(1))] = \sum_{c,w} E[Y | A=1, C=c, W=w] \times P[c,w]\]
9.3.3.3 \(\text{CE}_3\) estimation
Defining \(G^1\) as the random distribution of versions of treatment \(k^1 \in \mathcal{K}^{1}\), \(\text{CE}_3\) expresses as:
\[\begin{equation} \text{CE}_3 = E[Y(1, r(C)] - E[Y(1, G^1)] \text{ with } P[G^1=k^1_i]=\dfrac{1}{|\mathcal{K}^1_{PM}|}, \tag{9.8} \end{equation}\]\(|.|\) denoting the cardinality of the set. In this formula, \(E[Y(1, G^1)]\) can be derived from equation (9.5):
\[E[Y(1, G^1)] = \dfrac{1}{|\mathcal{K}^1_{PM}|} \times \sum_{c,k^1_i} E[Y | A=1, K^1=k^1_i, C=c] \times P[c]\]
9.3.4 Alternative estimation methods
For the sake of simplicity and brevity, we only detailed the standardization in previous sections. However, other popular candidate methods can be used. Estimators based on the inverse probability weighting (IPW) and targeted maximum likelihood estimation (TMLE) will also be computed in the following sections. IPW has the particularity of not trying to model the outcome but rather the process of assigning treatments. Its theoretical bases have been described in section 8.2.2.2 and the details of its adaptation to multiple versions of treatment is provided in appendix C.2.3.
The TMLE methods are of a different nature (Van der Laan and Rose 2011). They combine an outcome model and a treatment model in order to obtain a doubly robust estimate, i.e., an estimate that is robust to a possible misspecification of either model. Moreover, the estimation is done in several steps in order to optimize the equilibrium bias-variance, not for the overall distribution of the data but specifically for the causal effect of interest. These methods also have the particularity of being very often used with machine learning algorithms to fit the outcome or treatment models, instead of the parametric models classically used in standardization and IPW methods. A more detailed description of TMLE properties and the choices that have been made to adapt it to the problem of precision medicine are available in appendix C.2.4.
9.3.5 Code
The methods detailed above have been implemented in R and applied to simulate data and PDX data. The code is provided in the form of R notebooks (simulated data and PDX data) as well as in the form of an RShiny interactive application (simulated data only). All of these files are available in the dedicated GitHub repository.
9.4 Application to simulated data
The proposed methods are first tested on simulated data in order to check the performance of the estimators in finite sample sizes.
9.4.1 General settings
Response variable | Intercept | Lin. coeff. \(Y \sim C_1\) | Lin. coeff. \(Y \sim C_2\) |
---|---|---|---|
\(Y(0,k^0)\) | 0 | 0 | 15 |
\(Y(1,k^1_1)\) | -25 | -15 | 10 |
\(Y(1,k^1_2)\) | 0 | 0 | -20 |
Using the R package lava based on latent variable models, we simulate a super-population of 10,000 patients with variables \(C\), \(A\), \(K\) and \(Y\) as in Figure 9.5. We first define two independent binary variables \(C_1\) and \(C_2\), representing mutational status of genomic covariates, with a mutation prevalence of 40%. By analogy with the PDX data, \(Y\) represents the evolution of tumor volume and a low value (a fortiori negative) corresponds to a better response. Y is therefore defined as a continuous gaussian variable. For each counterfactual variable of response \(Y(a, k^a)\), we specify the intercept and the linear regression coefficients regarding influence of \(C_i\) as described in Table 9.1. Lower intercepts correspond to better responses/more efficient drugs. Similarly, a negative regression coefficient between \(Y(a, k^a_i)\) and \(C_j\) means that the gene \(C_j\) improves the response to \(k^a_i\). So all in all, \(k^1_1\) has the best basal response (lowest intercept). \(C_1\) (resp. \(C_2\)) improves the response to \(k^1_1\) (resp. \(k^1_2\)). The treatment algorithm of precision medicine is in line with these settings since patients mutated in \(C_1\) (regardless their \(C_2\) status) are recommended to take \(k^1_1\) and patients mutated for \(C_2\) only are recommended to take \(k^1_2\). Patients without mutations are not eligible for precision medicine and not taken into account in the computations. Since \(k^1_1\) has the best basal response we assume it is assigned with greater probability by the physician and implement the following distribution of observed treatments:
\[P[K=k^1_1]=0.5 \text{ and } P[K=k^1_2]=P[K=k^0]=0.25\]
A super-population of 10,000 patients is then generated. 1,000 cohorts of 200 patients are sampled without replacement within this super-population which, with the prevalences defined for the mutations, corresponds to an effective sample size of about 130 patients eligible for the PM algortithm. The causal effects \(\text{CE}_1\), \(\text{CE}_2\) and \(\text{CE}_3\) are computed based on different methods on the sub-cohort eligible for precision medicine (Figure 9.6):
- True effects, using all simulated counterfactuals for all patients
- Naive effects, using observed outcomes only for both PM and control arms
- Corrected effects: using observed outcome and standardized estimators (Std), inverse probability weighting (IPW) and targeted maximum likelihood estimators (TMLE).
9.4.2 Simulation results
First, the distribution of data in the super-population of 10,000 patients can be observed in Figure 9.7A, illustrating the different relations and differences described above. In particular, \(Y(1, k^1_1)\) (resp. \(Y(1, k^1_2)\)) is lower for \(C_1\)-mutated (resp. \(C_2\)-mutated) patients. It can also be seen that the response to precision medicine (\(Y(1, r(C))\)) differs according to the groups: patients mutated for \(C_1\) only have the best response, followed by patients mutated for both \(C_1\) and \(C_2\) and patients mutated for \(C_2\) only. There is therefore a heterogeneity of responses to PM which encourages to take into account the groups of patients and their PM versions. The right side of Figure 9.7A shows the deterministic assignment of the recommended PM treatment (\(r(C)\)) to each patient profile and the unbalanced distribution of observed treatments (\(K\)) with a predominance of \(k^1_1\).
In the first target trial, true \(\text{CE}_1\) estimates in the sampled cohorts are distributed around -40 (Figure 9.7B), confirming the superiority of the PM arm over the control arm as defined in the simulation parameters. Not all methods of estimating the causal effect perform equally well. The so-called naive estimate and the one based on IPW show a net bias. The over-representation of the most advantaged patients by PM tends to cause these methods to overestimate the benefit of PM, as can also be seen in the deviation plots. The same trends are observed for \(\text{CE}_2\) and \(\text{CE}_3\) (Figure 9.7C and D) where the differences are even more drastic. The mean absolute error of the naive method is thus divided by more than 2 when using standardized estimates or the TMLE.
In order to further dissect the influence of simulation parameters on estimation performances, a slightly different simulation scenario with equal probabilities of observed treatments has been studied: \[P[K=k^0]=P[K=k^1_1]=P[K=k^1_2]=\dfrac{1}{3}\] In this case, the random and balanced assignment of the observed treatments logically removes the systematic biases of the naive method by providing them with more randomized data. However, the corrections made by the proposed methods of causal inference, and in particular standardization and TMLE, still reduce the variances in the estimates due to the heterogeneity of the effects of precision medicine as a function of molecular profiles. Randomly, some sampled cohorts are indeed found with an association between \(C\) and observed \(K\), thus generating a confounding effect that the causal methods partially correct.
The simulated data allow us to imagine an almost infinite number of scenarios depending on the number of biomarkers taken into account in the algorithm, the number of different treatments, the dependencies of their responses or the distribution of treatments observed. In order to allow easy exploration of these scenarios without having to master the underlying R code, an interactive RShiny application has been developped. It can be accessed by locally running the R source file or by using the online version embedded in Figure 9.8. Readers with the ability to run the application locally are encouraged to favor this option because the hosting of the online application is limited to a maximum amount of time per month. The application allows certain additional analyses not presented in this manuscript, in particular the linking of biases observed in the sampled cohorts according to their composition (prevalence of mutations, treatments, etc.). It is thus possible to trace the origin of the biases.
9.5 Application to PDX
The method is then applied to public data from patient-derived xenografts (Gao et al. 2015), described in section 9.1.1 and appendix A.2. One of the major interests of this type of data in the context of this chapter is to provide access to treatment response values otherwise considered as hypothetical (or counterfactual). It is indeed possible to have the response of the same tumor (or more precisely of distinct samples from the same tumor) to different treatments, thus representing proxies for counterfactual variables, as described in Figure 9.1. Availability of these data provides a unique ground truth to assess the validity of proposed causal estimates in a pre-clinical context.
Based on the analysis accompanying the published data (Gao et al. 2015), some biomarkers of treatment response have been selected and resulted in an example of treatment algorithm: binimetinib (MEK inhibitor) is recommended to KRAS/BRAF mutated tumors, and BYL719 (alpha-specific PI3K inhibitor, also known as Alpelisib) to PIK3CA mutated tumors. PTEN is also included as a covariate because of its detrimental impact on the response to these two treatments. LEE011 drug (a cell cycle inhibitor also known as Ribociclib) is chosen as the reference drug treatment (\(k^0\)). Among the sequenced tumors, 88 are eligible for this precision medicine algorithm (i.e., mutated for BRAF, KRAS or PIK3CA) and have been tested for all 3 drugs of interest, thus ensuring the availability of all corresponding responses. The following analyses will focus exclusively on this sub-cohort for which a descriptive analysis is provided in Figure 9.9A. As expected BRAF/KRAS-mutated tumors have a better response to binimetinib and PIK3CA-mutated tumors have a better response to BYL719 (Figure 9.9B). In addition, it can be noted that these biomarkers have deleterious cross-effects.
The analysis settings are similar to the ones used for simulated data. 1,000 different cohorts of 70 tumors (out of 88) are sampled without replacement assuming each time that only the response to one of the treatments is known for each tumor, reproducing the classical clinical situation. The distribution of the observed treatments was defined randomly: \[P[K=k^0]=P[K=k^1_1]=P[K=k^1_2]=\dfrac{1}{3}\] It should be noted that, contrary to analyses based on simulated data, all the statistical models used for standardization (outcome model), for the IPW (treatment model) and for the TMLE are no longer generalized linear models (GLM) but random forests (RF). Indeed, it was observed that the performance of GLM-based methods was lower than that of the naive method, supporting the importance of relevant model specification consistent with real data. The RF algorithms then allow to limit misspecification due to the largely non-linear nature of the data. Random forests were chosen for their speed and versatility, especially in view of their ability to handle multinomial classification as well.
The results of estimations are then presented in Figure 9.10. In the presence of randomly assigned and balanced observed treatments, none of the methods (including the naive one) has significant systematic bias. On the other hand, more sophisticated methods, and in particular TMLE, allow to reduce the gap between estimates and true values, as visible on the mean absolute errors in Figure 9.10 right column. An additional analysis using the binary version of outcome \(Y\) is presented in Béal and Latouche (2020) with similar trends and conclusions: standardized, IPW and TMLE estimates are closer than naive methods to the true values from PDX. It supports the validity of the extension of the method to binary outcomes. In the same way as before with the simulated data, it would be possible to study the impact of non-random assignment of the observed treatments, which could systematically bias the results of the naive methods.
9.6 Limitations and perspectives
In synthesis, this work proposes a conceptual framework for evaluating a precision medicine algorithm, taking advantage of data already generated using adapted causal inference tools. However, in a clinical context, these data were not generated in a purely observational manner. Patients were cared for and treated by physicians who probably took into account some of their characteristics. However, the reasoning, formalized or not, behind the physicians' decisions does not correspond to that which a new investigator might want to test. In the eyes of this new investigator, the data can therefore be considered as observational in that they do not correspond to the randomization he would have liked to have carried out. The possibility for this new investigator to estimate the impact of his PM algorithm using the proposed estimators depends, however, on the consistency, exchangeability and positivity hypotheses.
The hypothesis of consistency has been made more plausible by taking into account the treatment versions, which makes it possible to explicit the heterogeneity of the molecules administered. Exchangeability remains questionable. The simulations and calculations described above underline the importance of taking into account at least the genomic covariates used in the processing algorithm. The inclusion of additional covariates is likely to be necessary in many real-world applications. Positivity, on the other hand, can be violated in a much more obvious way in certain situations. Thus, equation (9.5) requires positivity to be extended to versions of treatment: \(0<P[A=a, K^a=k^a|C]<1\). If the assignment of the observed treatments was done on a deterministic basis with respect to the variables used by the treatment algorithm, each patient's molecular profile will have been treated with a single drug, thus preventing any subsequent causal inference within the defined framework. The eventual use, by the boards of physicians in charge of assigning the observed treatments, of variables different from those used by the algorithm could then make it possible to verify the positive condition. But these variables would represent unmeasured confounding factors. It is therefore essential to have an in-depth knowledge of the rationales at work in the assignment of the observed treatments.
We developed a user-friendly application that extends the scope of the simulations and makes possible to study and quantify the impact of different situations, including possible (quasi-)violations of positivity or unmeasured confounding. It is thus a tool for empirically framing cases where this causal inference is reasonable or not. The analysis of the PDX data provides an illustration and proof of feasibility for these methods on pre-clinical data, closer to the human clinical data generally of interest. Beyond feasibility, this implementation leads to some remarks. Firstly, the improvement of causal inference methods compared to naive estimation of PM effects is conditioned in this case to the use of flexible and non-linear learning algorithms. This underlines the importance of a proper specification of the outcome and treatment models whose imperfection, especially when trained on small samples, could explain the modesty of the results compared to the simulated data. The particular nature of the PDX data design used should also be kept in mind: each tumor is tested only once for each drug, which may lead to greater variability of results due to tumor heterogeneity (Gao et al. 2015). Some studies, with smaller numbers of tumors and treatments, propose to form groups of several mice for each treatment-drug combination (Hidalgo et al. 2014). The use of these mean effects could contribute to more accurate data. In spite of these limitations, which may diminish their ability to provide values with counterfactual interpretation, PDX data are thus a dataset of interest for studying and validating methods of causal inferences about treatment response. It can also be noted that the very nature of these data, due to the multiplicity of drugs tested for each tumor, can provide a framework in which the constraints of positivity are singularly alleviated. Even if all drugs were not tested on all patients, considering each tumor-drug combination as a different unit increases the coverage of the data. It is then necessary to take into account the clustered nature of the data, each tumor being present several times.
Finally, beyond the pre-clinical data presented here, the theoretical framework developed in this chapter should be more directly applicable to data from clinical trials if these data do not violate the requirements of positivity. If it is necessary to consider several trials, the heterogeneity of practices must be taken into account. The use of different drug lists from one trial to another or from one medical centre to another could also provide an example of confounding factor \(W\), included in the theoretical framework presented here but not used in applications.
Summary
The emergence of targeted treatments coupled with a better understanding of patients' molecular profiles has fostered the development of precision medicine for cancer: the most appropriate treatment is defined according to the patient's genetic alterations. Because they can in some cases allow the simulation and comparison of different treatments for a given molecular profile, mechanistic models can thus be considered as an example of those treatment assignment algorithms that represent precision medicine. It is therefore desirable to evaluate the clinical benefits of these algorithms using clinical trials or causal inference. In the second case explored here, the theoretical framework is extended to allow the existence of different drugs, or versions of treatment, in the treated patients. Different statistical estimators are defined to emulate several clinical trials and estimate the corresponding causal effect of precision medicine algorithms. These methods are tested on simulated data according to different scenarios and an interactive application is proposed to explore others. A second application is presented on preclinical data from PDX, which provides for each tumor the experimental response to different treatments. Access to these data, otherwise considered as counterfactual, allows to validate the capacity of the developed methods to reduce bias in the estimation of effects.
References
Abou-Jawde, Rony, Toni Choueiri, Carlos Alemany, and Tarek Mekhail. 2003. “An Overview of Targeted Treatments in Cancer.” Clinical Therapeutics 25 (8). Elsevier: 2121–37.
Béal, Jonas, and Aurélien Latouche. 2020. “Causal Inference with Multiple Versions of Treatment and Application to Personalized Medicine.” ArXiv Preprint ArXiv:2005.12427.
Coyne, Geraldine O’Sullivan, Naoko Takebe, and Alice P Chen. 2017. “Defining Precision: The Precision Medicine Initiative Trials Nci-Mpact and Nci-Match.” Current Problems in Cancer 41 (3). Elsevier: 182–93.
Cree, Ian A, Christian M Kurbacher, Alan Lamont, Andrew C Hindley, Sharon Love, TCA Ovarian Cancer Trial Group, and others. 2007. “A Prospective Randomized Controlled Trial of Tumour Chemosensitivity Assay Directed Chemotherapy Versus Physician’s Choice in Patients with Recurrent Platinum-Resistant Ovarian Cancer.” Anti-Cancer Drugs 18 (9). LWW: 1093–1101.
De Gramont, Armand, Sarah Watson, Lee M Ellis, Jordi Rodón, Josep Tabernero, Aimery De Gramont, and Stanley R Hamilton. 2015. “Pragmatic Issues in Biomarker Evaluation for Targeted Therapies in Cancer.” Nature Reviews Clinical Oncology 12 (4). Nature Publishing Group: 197.
Flaherty, Keith T, Robert Gray, Alice Chen, Shuli Li, David Patton, Stanley R Hamilton, Paul M Williams, et al. 2020. “THE Molecular Analysis for Therapy Choice (Nci-Match) Trial: LESSONS for Genomic Trial Design.” JNCI: Journal of the National Cancer Institute.
Freidlin, Boris, Lisa M McShane, and Edward L Korn. 2010. “Randomized Clinical Trials with Biomarkers: Design Issues.” Journal of the National Cancer Institute 102 (3). Oxford University Press: 152–60.
Friedman, Adam A, Anthony Letai, David E Fisher, and Keith T Flaherty. 2015. “Precision Medicine for Cancer with Next-Generation Functional Diagnostics.” Nature Reviews Cancer 15 (12). Nature Publishing Group: 747–56.
Gao, Hui, Joshua M Korn, Stéphane Ferretti, John E Monahan, Youzhen Wang, Mallika Singh, Chao Zhang, et al. 2015. “High-Throughput Screening Using Patient-Derived Tumor Xenografts to Predict Clinical Trial Drug Response.” Nature Medicine 21 (11). Nature Publishing Group: 1318.
Hansen, J, and R Iyengar. 2013. “Computation as the Mechanistic Bridge Between Precision Medicine and Systems Therapeutics.” Clinical Pharmacology & Therapeutics 93 (1). Wiley Online Library: 117–28.
Hernán, Miguel A, and James M Robins. 2016. “Using Big Data to Emulate a Target Trial When a Randomized Trial Is Not Available.” American Journal of Epidemiology 183 (8). Oxford University Press: 758–64.
Hernán, Miguel A, and Tyler J VanderWeele. 2011. “Compound Treatments and Transportability of Causal Inference.” Epidemiology (Cambridge, Mass.) 22 (3). NIH Public Access: 368.
Hidalgo, Manuel, Frederic Amant, Andrew V Biankin, Eva Budinská, Annette T Byrne, Carlos Caldas, Robert B Clarke, et al. 2014. “Patient-Derived Xenograft Models: An Emerging Platform for Translational Cancer Research.” Cancer Discovery 4 (9). AACR: 998–1013.
Le Tourneau, Christophe, Jean-Pierre Delord, Anthony Gonçalves, Céline Gavoille, Coraline Dubot, Nicolas Isambert, Mario Campone, et al. 2015. “Molecularly Targeted Therapy Based on Tumour Molecular Profiling Versus Conventional Therapy for Advanced Cancer (Shiva): A Multicentre, Open-Label, Proof-of-Concept, Randomised, Controlled Phase 2 Trial.” The Lancet Oncology 16 (13). Elsevier: 1324–34.
Ma, Jianzhu, Michael Ku Yu, Samson Fong, Keiichiro Ono, Eric Sage, Barry Demchak, Roded Sharan, and Trey Ideker. 2018. “Using Deep Learning to Model the Hierarchical Structure and Function of a Cell.” Nature Methods 15 (4). Nature Publishing Group: 290.
Manica, Matteo, Ali Oskooei, Jannis Born, Vigneshwari Subramanian, Julio Sáez-Rodríguez, and María Rodríguez Martínez. 2019. “Toward Explainable Anticancer Compound Sensitivity Prediction via Multimodal Attention-Based Convolutional Encoders.” Molecular Pharmaceutics. ACS Publications.
Rodon, Jordi, Jean-Charles Soria, Raanan Berger, Wilson H Miller, Eitan Rubin, Aleksandra Kugel, Apostolia Tsimberidou, et al. 2019. “Genomic and Transcriptomic Profiling Expands Precision Cancer Medicine: The Winther Trial.” Nature Medicine 25 (5). Nature Publishing Group: 751.
Rubin, Donald B. 1980. “Randomization Analysis of Experimental Data: The Fisher Randomization Test Comment.” Journal of the American Statistical Association 75 (371). JSTOR: 591–93.
Sargent, Daniel J, Barbara A Conley, Carmen Allegra, and Laurence Collette. 2005. “Clinical Trial Designs for Predictive Marker Validation in Cancer Treatment Trials.” Journal of Clinical Oncology 23 (9). American Society of Clinical Oncology: 2020–7.
Van der Laan, Mark J, and Sherri Rose. 2011. Targeted Learning: Causal Inference for Observational and Experimental Data. Springer Science & Business Media.
VanderWeele, Tyler J. 2009. “Concerning the Consistency Assumption in Causal Inference.” Epidemiology 20 (6). LWW: 880–83.
VanderWeele, Tyler J, and Miguel A Hernan. 2013. “Causal Inference Under Multiple Versions of Treatment.” Journal of Causal Inference 1 (1). De Gruyter: 1–20.
Von Hoff, Daniel D, Joseph J Stephenson Jr, Peter Rosen, David M Loesch, Mitesh J Borad, Stephen Anthony, Gayle Jameson, et al. 2010. “Pilot Study Using Molecular Profiling of Patients’ Tumors to Find Potential Targets and Select Treatments for Their Refractory Cancers.” Journal of Clinical Oncology 28 (33). American Society of Clinical Oncology: 4877–83.