Core Concepts
Leveraging pretrained imputation models and fine-tuning them for downstream classification tasks in Electronic Health Records (EHRs) can achieve comparable performance to more complex architectures, even with simple classifiers.
Stats
The highest average AUC on the MIMIC III dataset was achieved by a 2-layered MLP, reaching 0.839 on the validation set and 0.819 on the test set.
For the Physionet 2012 dataset, the 5-layered MLP achieved the highest average AUCs, reaching 0.824 on the validation set and 0.805 on the test set.
The simplest model, the 2-layered MLP (14,081 parameters), achieved a test AUC of 0.797 and a validation AUC of 0.811 on the Physionet dataset.
The LSTM model had 76,721 parameters, and the GRU model had 61,061 parameters.
Quotes
"This paper explores the development of a modular, deep learning-based imputation and classification pipeline, specifically built to leverage the capabilities of state-of-the-art imputation models for downstream classification tasks."
"This is problematic because it raises questions about whether the improvements in the observed performances are due to the strength of the imputer model or the complexity of the downstream model or their interaction."
"Our research focuses on the combination of both approaches, where we freeze the weights of our foundational model and utilise the learned features as input to the classifier to perform downstream tasks. We also explore an additional approach, wherein the weights of the pretrained model aren’t frozen and are trained (fine-tuned) along with our downstream classifier."