66
views
0
recommends
+1 Recommend
0 collections
    0
    shares
      • Record: found
      • Abstract: found
      • Article: found
      Is Open Access

      Explicit causal reasoning is needed to prevent prognostic models being victims of their own success

      research-article

      Read this article at

      Bookmark
          There is no author summary for this article yet. Authors can add summaries to their articles on ScienceOpen to make them more accessible to a non-specialist audience.

          Abstract

          The recent perspective by Lenert et al 1 provides an accessible and informative overview of the full life cycle of prognostic models, comprising development, deployment, maintenance, and surveillance. The perspective focuses particularly on the fundamental issue that deployment of a prognostic model into clinical practice will lead to changes in decision making or interventions, and hence, changes in clinical outcomes. This has received little attention in the prognostic modeling literature but is important because this changes predictor-outcome associations, meaning that the performance of the model degrades over time; therefore, prognostic models become “victims of their own success.” More seriously, a prediction from such a model is challenging to interpret, as it implicitly reflects both the risk factors and the interventions that similar patients received, in the historical data used to develop the prognostic model. The authors rightly point out that “holistically modeling the outcome and interventions(s)” and “incorporat[ing] the intervention space” are required to overcome this concern. 1 However, the proposed solution of directly modeling interventions, or their surrogates, is not sufficient. An explicit causal inference framework is required. When the intended use of a prognostic model is to support decisions concerning intervention(s), the counterfactual causal framework provides a natural and powerful way to ensure that predictions issued by the prognostic model are useful, interpretable, and less vulnerable to degradation over time. The framework allows predictions to be used to answer “what if” questions; for an introduction, see Hernan and Robbins. 2 However, appropriate modeling of these counterfactual scenarios is far more challenging than pure prediction, particularly in the presence of time-dependent confounding. Here, standard regression modeling becomes inadequate and specialist techniques are required. 2 In the scenarios carefully articulated by Lenert et al, in which risk models are used to alert to a high-risk situation and thereby inform intervention, one should primarily be interested in the counterfactual “treatment-naïve” prediction: in other words, “what is the risk of outcome for this individual if we do not intervene?” Failure to explicitly model this treatment-naïve prediction will lead to high-risk patients being classified inappropriately as low risk, as their prediction is reflective of interventions made to lower the risk of similar patients in the past. 3 This situation becomes more pronounced when a successful model is updated, as interventions made based on the predictions from the model are hoped to change the risk. Recently, we illustrated how to calculate treatment-naïve risk in the presence of “treatment drop-in,” a scenario in which patients begin taking treatments after the time a prediction is made but before the outcome. 4 With treatment-naïve risk as a baseline, one can move to evaluating predictions under a range of different interventions; the counterfactual causal framework allows a model to be interrogated with a series of “what if” questions. Comparison of the outcome predictions or distributions under different scenarios can then naturally provide information to support intervention decisions. Alongside this counterfactual framework, we agree with Lenert et al that “robust performance surveillance of models in clinical use” is required postdeployment as part of prognostic model maintenance and model surveillance. However, doing this through so-called static updating, in which previous iterations of a risk model are refined according to new datasets observed in batches, still requires timely identification of performance drift. This often leads to an identification-action latency period, in which noticing and acting on a deterioration in a model’s performance occurs much later in time than should be acceptable in clinical practice. This is amplified by a lower frequency of updating but could be mitigated through continuous surveillance and maintenance of the prognostic models. So-called dynamic modeling is an emerging area of research 5 that enables the continuous incorporation of surveillance and refinement directly into the modeling processes and could prevent prognostic models being “victims of their own success” if combined appropriately with counterfactual frameworks. While counterfactual prediction is only beginning to be applied in prognostic model development, it is a technique that will allow many of the issues eloquently described by Lenert and colleagues to be mitigated. Moreover, it provides predictions that are arguably closer to what a decision maker needs, and likely to be more robust over time. CONFLICT OF INTEREST STATEMENT None declared.

          Related collections

          Most cited references4

          • Record: found
          • Abstract: found
          • Article: found
          Is Open Access

          Dynamic models to predict health outcomes: current status and methodological challenges

          Background Disease populations, clinical practice, and healthcare systems are constantly evolving. This can result in clinical prediction models quickly becoming outdated and less accurate over time. A potential solution is to develop ‘dynamic’ prediction models capable of retaining accuracy by evolving over time in response to observed changes. Our aim was to review the literature in this area to understand the current state-of-the-art in dynamic prediction modelling and identify unresolved methodological challenges. Methods MEDLINE, Embase and Web of Science were searched for papers which used or developed dynamic clinical prediction models. Information was extracted on methods for model updating, choice of update windows and decay factors and validation of models. We also extracted reported limitations of methods and recommendations for future research. Results We identified eleven papers that discussed seven dynamic clinical prediction modelling methods which split into three categories. The first category uses frequentist methods to update models in discrete steps, the second uses Bayesian methods for continuous updating and the third, based on varying coefficients, explicitly describes the relationship between predictors and outcome variable as a function of calendar time. These methods have been applied to a limited number of healthcare problems, and few empirical comparisons between them have been made. Conclusion Dynamic prediction models are not well established but they overcome one of the major issues with static clinical prediction models, calibration drift. However, there are challenges in choosing decay factors and in dealing with sudden changes. The validation of dynamic prediction models is still largely unexplored terrain.
            Bookmark
            • Record: found
            • Abstract: found
            • Article: not found

            Prognostic models will be victims of their own success, unless…

            Predictive analytics have begun to change the workflows of healthcare by giving insight into our future health. Deploying prognostic models into clinical workflows should change behavior and motivate interventions that affect outcomes. As users respond to model predictions, downstream characteristics of the data, including the distribution of the outcome, may change. The ever-changing nature of healthcare necessitates maintenance of prognostic models to ensure their longevity. The more effective a model and intervention(s) are at improving outcomes, the faster a model will appear to degrade. Improving outcomes can disrupt the association between the model’s predictors and the outcome. Model refitting may not always be the most effective response to these challenges. These problems will need to be mitigated by systematically incorporating interventions into prognostic models and by maintaining robust performance surveillance of models in clinical use. Holistically modeling the outcome and intervention(s) can lead to resilience to future compromises in performance.
              Bookmark
              • Record: found
              • Abstract: found
              • Article: found
              Is Open Access

              Using marginal structural models to adjust for treatment drop‐in when developing clinical prediction models

              Clinical prediction models (CPMs) can inform decision making about treatment initiation, which requires predicted risks assuming no treatment is given. However, this is challenging since CPMs are usually derived using data sets where patients received treatment, often initiated postbaseline as “treatment drop‐ins.” This study proposes the use of marginal structural models (MSMs) to adjust for treatment drop‐in. We illustrate the use of MSMs in the CPM framework through simulation studies that represent randomized controlled trials and real‐world observational data and the example of statin initiation for cardiovascular disease prevention. The simulations include a binary treatment and a covariate, each recorded at two timepoints and having a prognostic effect on a binary outcome. The bias in predicted risk was examined in a model ignoring treatment, a model fitted on treatment‐naïve patients (at baseline), a model including baseline treatment, and the MSM. In all simulation scenarios, all models except the MSM underestimated the risk of outcome given absence of treatment. These results were supported in the statin initiation example, which showed that ignoring statin initiation postbaseline resulted in models that significantly underestimated the risk of a cardiovascular disease event occurring within 10 years. Consequently, CPMs that do not acknowledge treatment drop‐in can lead to underallocation of treatment. In conclusion, when developing CPMs to predict treatment‐naïve risk, researchers should consider using MSMs to adjust for treatment drop‐in, and also seek to exploit the ability of MSMs to allow estimation of individual treatment effects.
                Bookmark

                Author and article information

                Journal
                J Am Med Inform Assoc
                J Am Med Inform Assoc
                jamia
                Journal of the American Medical Informatics Association : JAMIA
                Oxford University Press
                1067-5027
                1527-974X
                December 2019
                14 November 2019
                14 November 2019
                : 26
                : 12
                : 1675-1676
                Affiliations
                Division of Informatics, Imaging and Data Science, Faculty of Biology, Medicine and Health, University of Manchester, Manchester, United Kingdom
                Author notes
                Corresponding Author: Matthew Sperrin, PhD, Vaughan House, Portsmouth Street, University of Manchester, Manchester M13 9PL, UK; matthew.sperrin@ 123456manchester.ac.uk
                Article
                ocz197
                10.1093/jamia/ocz197
                6857504
                31722385
                4bf5e829-bc2e-432e-9201-9544db68a713
                © The Author(s) 2019. Published by Oxford University Press on behalf of the American Medical Informatics Association.

                This is an Open Access article distributed under the terms of the Creative Commons Attribution Non-Commercial License ( http://creativecommons.org/licenses/by-nc/4.0/), which permits non-commercial re-use, distribution, and reproduction in any medium, provided the original work is properly cited. For commercial re-use, please contact journals.permissions@oup.com

                History
                : 11 September 2019
                : 18 October 2019
                Page count
                Pages: 2
                Categories
                Correspondence

                Bioinformatics & Computational biology
                Bioinformatics & Computational biology

                Comments

                Comment on this article