Conclusion
One path among others
The aim of this thesis was to trace a path to link the biological knowledge of cancer to the clinical impact through mechanistic models. At each of the stages that separate the creation of a mechanistic model of cancer from its possible clinical application, choices were made that determined the obstacles encountered and the final shape of this work. Before elaborating on the subject, it is worth summarizing one last time the thread followed during this thesis.
Among the many possible orientations, it was chosen to take the opposite side of the data-intensive machine learning methods. The main approach proposed uses a qualitative logical formalism and integrates the data by interpreting them rather than by optimizing the parameters with respect to a particular objective. As a result, the resulting personalized mechanistic models have proven to be more of an interpretive than a predictive tool. Their versatility and low data requirements nevertheless allow them to be applied to a wide range of questions, particularly concerning the response to treatments that their mechanistic nature facilitates. This seemingly limitless versatility can, however, prove to be a trap because, while all kinds of applications are theoretically possible, the need to rely on detailed biological knowledge and appropriate data limits its scope.
In the case of mechanistic molecular signaling models, this interpretive nature of the models is confirmed by statistical analyses. The main value of these models is to provide an understandable framework for extracting relevant biological information in the context of current biological knowledge. The ability of these models to detect emerging non-linear information is also proven, but is rarer and of relatively smaller magnitude. Given the influx of biological knowledge and data, computational models of cancer, with various formalisms, are nevertheless multiplying, particularly with medical aims. In the context of cancer, their use to recommend personalised treatment for each patient is a possible horizon. The evaluation of these models could then become increasingly acute. This thesis proposes the adaptation of causal inference methods in order to simulate their evaluation in clinical trials and to come as close as possible to medical evaluation standards.
In other directions
Along this path, many forks are conceivable, as well as the further pursuit of some of the paths that have been explored. Without going back over the technical limits and perspectives of parts II and III, already mentioned in sections 6.4 and 9.6 respectively, it is possible to paint a broader picture of some of the obstacles and alternatives.
If we question the different choices and presuppositions of this thesis we must first mention that mechanistic models of cancer are not limited to the modeling of molecular networks. Other fields have been explored, and some applications are already more mature and closer to direct use in clinical oncology. This is essentially the case for pharmacological approaches focusing on the pharmacokinetics or pharmacodynamics of certain compounds and the modeling of tumor size (Benzekry 2020). The two approaches answer different questions from those raised in this thesis, but they are nevertheless closely linked and we must consider the enrichments they can bring. For example, pharmacokinetic and pharmacodynamic modeling represents a very clear blind spot of the approaches presented in chapter 6 where all BRAF inhibitors are modeled in an identical and crude way: quantitative parameters of BRAF inhibition are not translated. All these aspects constitute the core of mechanistic pharmacokinetic/pharmacodynamic models that are biochemically more accurate and fine-grained. On the other hand, tumor size models present a more macroscopic approach that is easier to relate to patient-level clinical interpretations.
All these considerations thus incite us not to consider a single-layer model and to try to make it cover the entire path, from the molecule to the patient, as proposed in this thesis. Rather than building huge mechanistic models covering all scales, it seems appropriate to work on linking different models in a modular multi-scale approach making it easier to cooperate between fields of modeling that often require different knowledge. As an example, it is possible to combine MaBoSS and its molecular network models with PhysiCell, a 3D multicellular system simulator (Ghaffarizadeh et al. 2018). The PhysiBoSS framework schematically proposes to encapsulate a logical model in each cell to simulate its internal functions and then let PhysiCell model the three-dimensional interactions between multiple cells of potentially different types (Letort et al. 2019). Probably the most promising prospects for mechanistic models lie in synergies of this kind, which, however, require broader collaborations and longer design times to truly enable the different modeling scales to communicate. The idea of proposing multi-scale models encapsulated in a hierarchical way to ensure the modeling from the genetic scale to the phenotype is also found in intermediate approaches between mechanistic and deep learning models that force neural networks to follow this architecture in order to ease post hoc interpretation (Ma et al. 2018).
The previous mention of quantitative pharmacology models also underlines the intrinsic limitations of qualitative mechanistic models, a fortiori for clinical applications. Thus the outputs of the personalized logical models can only be interpreted in relation to each other; it is difficult to give each one an absolute interpretation. Without proper calibration, what does a Proliferation score of 0.2 or 0.8 mean? On first reading, they have only a relative significance that precludes direct clinical interpretation. However, it seems difficult to make these logical models much more quantitative. The very nature of formalism limits the precise translation of biological mechanisms retrieved from literature. Likewise, the partial integration of some omics data only in logical models, presented in chapter 5, required a specific methodology to allow the data to be adapted to the discrete and qualitative formalism. Obtaining truly quantitative and well-calibrated mechanistic models has until now always required intrinsically more quantitative formalisms such as ordinary differential equations (Fröhlich et al. 2018), or Bayesian frameworks allowing a more flexible representation of entities and their relationships (Jastrzebski et al. 2018). In these two cases, as in most others, the quantitative quality of the predictions relies on model training and parameter fitting, contrary to the approach proposed in this thesis.
These methods therefore depend very heavily on the available data to train their models. One of their limitations is to use data from cell line screening. Schematically it is necessary to have the response to different treatments for each line in order to obtain personalized models or interpretations. The abundance of this type of screening data makes them the preferred methods for these applications where they greatly outperform qualitative models without learning. On the other hand, in a more prospective way, it is interesting to ask the question of the availability of patient data with the view to one day apply these models as close as possible to the clinic. Today it is inconceivable to obtain such rich data for patients. This would imply being able to develop cell lines for each patient, then screen all of them for drugs and use the results. This is a prospect whose feasibility is being studied (Lee et al. 2018), but it is not a routine clinical procedure today. If it were possible, the models would probably be replaced by the experiments themselves. More realistically, it is hoped that once the model has been sufficiently trained, the data required to make predictions for a new patient will be more modest. However, this new patient still needs to remain within the training data field. Conversely, logical models personalized without learning are based on other assumptions, mainly related to the correct mathematical translation of biological knowledge and interpreted omics data. Continued growth in the quantity and quality of data is likely to give an overall advantage in the medium term to the above mentioned quantitative mechanistic methods, more data-intensive but more accurate. Qualitative models such as those presented in chapters 5 and 6 may nonetheless remain a complementary approach, used beforehand or in more data-poor cases that do not allow for learning, as discussed when comparing the results with machine learning approaches in section 6.2.5.
In such a world of abundant data, is it only then relevant to continue to propose mechanistic models instead of biologically agnostic machine learning models? The latter have indeed become much more frequent with a particular emphasis in recent years on deep learning approaches (Angermueller et al. 2016; Eraslan et al. 2019). The comparison between quantitative mechanistic approaches and machine learning approaches then uses terms similar to the previous comparison between qualitative and quantitative mechanistic approaches. Machine learning approaches can be even more flexible because their degrees of freedom are not constrained by an a priori biological structure, but they consequently require larger amounts of data. The loss of any internal mechanistic structure can however hamper the interpretability of the results, although many methods have been developed to facilitate this interpretation of machine learning models in biology (Azuaje 2019; Manica et al. 2019). The loss of the mechanistic nature in most machine learning approaches also makes it impossible to test perturbations or hypotheses that are clearly outside the scope of training data. If one can imagine the predominance of well-calibrated machine learning models for routine decisions, more exploratory and prospective investigations could however favour mechanistic models more capable of extrapolation. Moreover, while the question of the acceptability of computational decision support models in oncology arises (Vollmer et al. 2020), the explicit representation of the internal mechanisms may present a reassuring aspect. It is simpler for both the patient and the physician to imagine the functioning of a model that replicates human bodies or cells than a more exclusively mathematical black box. However, this argument only seems valid in the case of very similar performances.
Finally, most of the remaining questions that arise concern all computational models of cancer in the same way, whatever their nature. Wiens et al. (2019) thus propose a roadmap for the responsible use of the models in the clinic that targets a number of challenges. One of them is to know how computational models can be seamlessly integrated into medical practice while leaving the various stakeholders, from medical experts to patients, in the loop. Physicians, for example, need to be informed in a relevant way about the nature and limitations of these models, which can facilitate the decision but whose technical details may escape them. The presentation of these models to them is therefore particularly crucial in order to ensure their proper use, as Sendak et al. (2020) point out by proposing standardized methods of presentation and communication. Numerous ethical or data privacy issues also arise when precise genetic information about patients needs to be processed.
One of the challenges mentioned by Wiens et al. (2019) which resonates particularly with this thesis is that of the rigorous evaluation of the clinical contribution of these models, mechanistic or not. The choice of metrics is therefore very important and must be made to optimize clinical information. For example, while the discrimination capability of models is frequently measured, the calibration performance of risk prediction models is also crucial to inform clinical practice (Van Calster et al. 2019). In general, although computational cancer models are changing the cancer research landscape, proven standards from epidemiology or clinical statistics for evaluation should be maintained to allow comparison and quantification of improvements beyond scientific trends (Christodoulou et al. 2019). These evaluations require real clinical expertise to measure the extent to which a gain in algorithmic performance can be translated into clinical improvement depending on the patients concerned and the resources mobilized. Concerning the mechanistic models in particular, for example, their capacity to extrapolate and possibly reason about new treatments for which data are missing or few have been underlined several times. However, the clinical evaluation of their impact always requires data at the end of the day. The methods of causal inference applied in chapter 9 make it possible to correct the effects and bring them closer to the standards of clinical trials, but they can only rearrange clinical strategies on the basis of existing data. Mechanistic models are therefore there to guide or enlighten experiments, but are not intended to replace them.
To conclude in a word, cancer models still have a bright future ahead of them. Mechanistic models will continue to be attractive because of their ability not only to predict but, more importantly, to explain. However, the transparency of their mechanisms should not prevent them from being rigorously evaluated statistically. It is not enough for them to explain, they must also be well understood.
References
Angermueller, Christof, Tanel Pärnamaa, Leopold Parts, and Oliver Stegle. 2016. “Deep Learning for Computational Biology.” Molecular Systems Biology 12 (7): 878.
Azuaje, Francisco. 2019. “Artificial Intelligence for Precision Oncology: Beyond Patient Stratification.” NPJ Precision Oncology 3 (1). Nature Publishing Group: 1–5.
Benzekry, Sébastien. 2020. “Artificial Intelligence and Mechanistic Modeling for Clinical Decision Making in Oncology.” Clinical Pharmacology & Therapeutics. Wiley Online Library.
Christodoulou, Evangelia, Jie Ma, Gary S Collins, Ewout W Steyerberg, Jan Y Verbakel, and Ben Van Calster. 2019. “A Systematic Review Shows No Performance Benefit of Machine Learning over Logistic Regression for Clinical Prediction Models.” Journal of Clinical Epidemiology 110. Elsevier: 12–22.
Eraslan, Gökcen, Žiga Avsec, Julien Gagneur, and Fabian J Theis. 2019. “Deep Learning: New Computational Modelling Techniques for Genomics.” Nature Reviews Genetics 20 (7). Nature Publishing Group: 389–403.
Fröhlich, Fabian, Thomas Kessler, Daniel Weindl, Alexey Shadrin, Leonard Schmiester, Hendrik Hache, Artur Muradyan, et al. 2018. “Efficient Parameter Estimation Enables the Prediction of Drug Response Using a Mechanistic Pan-Cancer Pathway Model.” Cell Systems 7 (6). Elsevier: 567–79.
Ghaffarizadeh, Ahmadreza, Randy Heiland, Samuel H Friedman, Shannon M Mumenthaler, and Paul Macklin. 2018. “PhysiCell: An Open Source Physics-Based Cell Simulator for 3-d Multicellular Systems.” PLoS Computational Biology 14 (2). Public Library of Science: e1005991.
Jastrzebski, Katarzyna, Bram Thijssen, Roelof JC Kluin, Klaas de Lint, Ian J Majewski, Roderick L Beijersbergen, and Lodewyk FA Wessels. 2018. “Integrative Modeling Identifies Key Determinants of Inhibitor Sensitivity in Breast Cancer Cell Lines.” Cancer Research 78 (15). AACR: 4396–4410.
Lee, Jin-Ku, Zhaoqi Liu, Jason K Sa, Sang Shin, Jiguang Wang, Mykola Bordyuh, Hee Jin Cho, et al. 2018. “Pharmacogenomic Landscape of Patient-Derived Tumor Cells Informs Precision Oncology Therapy.” Nature Genetics 50 (10). Nature Publishing Group: 1399–1411.
Letort, Gaelle, Arnau Montagud, Gautier Stoll, Randy Heiland, Emmanuel Barillot, Paul Macklin, Andrei Zinovyev, and Laurence Calzone. 2019. “PhysiBoSS: A Multi-Scale Agent-Based Modelling Framework Integrating Physical Dimension and Cell Signalling.” Bioinformatics 35 (7). Oxford University Press: 1188–96.
Ma, Jianzhu, Michael Ku Yu, Samson Fong, Keiichiro Ono, Eric Sage, Barry Demchak, Roded Sharan, and Trey Ideker. 2018. “Using Deep Learning to Model the Hierarchical Structure and Function of a Cell.” Nature Methods 15 (4). Nature Publishing Group: 290.
Manica, Matteo, Ali Oskooei, Jannis Born, Vigneshwari Subramanian, Julio Sáez-Rodríguez, and María Rodríguez Martínez. 2019. “Toward Explainable Anticancer Compound Sensitivity Prediction via Multimodal Attention-Based Convolutional Encoders.” Molecular Pharmaceutics. ACS Publications.
Sendak, Mark P, Michael Gao, Nathan Brajer, and Suresh Balu. 2020. “Presenting Machine Learning Model Information to Clinical End Users with Model Facts Labels.” NPJ Digital Medicine 3 (1). Nature Publishing Group: 1–4.
Van Calster, Ben, David J McLernon, Maarten Van Smeden, Laure Wynants, and Ewout W Steyerberg. 2019. “Calibration: The Achilles Heel of Predictive Analytics.” BMC Medicine 17 (1). BioMed Central: 1–7.
Vollmer, Sebastian, Bilal A Mateen, Gergo Bohner, Franz J Király, Rayid Ghani, Pall Jonsson, Sarah Cumbers, et al. 2020. “Machine Learning and Artificial Intelligence Research for Patient Benefit: 20 Critical Questions on Transparency, Replicability, Ethics, and Effectiveness.” Bmj 368. British Medical Journal Publishing Group.
Wiens, Jenna, Suchi Saria, Mark Sendak, Marzyeh Ghassemi, Vincent X Liu, Finale Doshi-Velez, Kenneth Jung, et al. 2019. “Do No Harm: A Roadmap for Responsible Machine Learning for Health Care.” Nature Medicine 25 (9). Nature Publishing Group: 1337–40.