Learning-Based Radiomic Prediction of Type 2 Diabetes Mellitus Using Image-Derived Phenotypes

Michael S. Yao, 1, 2 , Allison Chae, 2 , Matthew T. MacLean3, Anurag Verma4, Jeffrey Duda3, James Gee3, Drew A. Torigian3, Daniel Rader4, Charles Khan2, 3, Walter R. Witschey, 2, 3 & Hersh Sagreiya, 2, 3, *

1Department of Bioengineering, University of Pennsylvania

2School of Medicine, University of Pennsylvania

3Department of Radiology, University of Pennsylvania

4Department of Medicine, University of Pennsylvania

Equal contribution. {michael.yao, jisoo.chae}@pennmedicine.upenn.edu

Equal contribution. {witschey, hersh.sagreiya}@pennmedicine.upenn.edu

*Corresponding Author.

Project Links


Early diagnosis of Type 2 Diabetes Mellitus (T2DM) is crucial to enable timely therapeutic interventions and lifestyle modifications. As medical imaging data become more widely available for many patient populations, we sought to investigate whether image-derived phenotypic data could be leveraged in tabular learning classifier models to predict T2DM incidence without the use of invasive blood lab measurements. We show that both neural network and decision tree models that use image-derived phenotypes can predict patient T2DM status with recall scores as high as 87.6%. We also propose the novel use of these same architectures as 'SynthA1c encoders' that are able to output interpretable values mimicking blood hemoglobin A1C empirical lab measurements. Finally, we demonstrate that T2DM risk prediction model sensitivity to small perturbations in input vector components can be used to predict performance on covariates sampled from previously unseen patient populations.


  title={Learning-Based Radiomic Prediction of {Type 2 Diabetes Mellitus} Using Image-Derived Phenotypes},
  authors={Yao, Michael S and Chae, Allison and MacLean, Matthew T and Verma, Anurag and Duda, Jeffrey and Gee, James and Torigian, Drew A and Rader, Daniel and Khan, Charles and Witschey, Walter R and Sagreiya, Hersh},