Adversarial Masking for Pretraining ECG Data Improves Downstream Model GeneralizabilityDownload PDF

Published: 02 Dec 2022, Last Modified: 05 May 2023TS4H PosterReaders: Everyone
Keywords: Data augmentation, self-supervised learning, ECG data
TL;DR: We adversarially generate masks as augmentations for 12-lead ECG data during contrastive self-supervised pretraining, which improves downstream task performance over traditional time-series data augmentations.
Abstract: Medical datasets often face the problem of data scarcity, as ground truth labels must be generated by medical professionals. One mitigation strategy is to pretrain deep learning models on large, unlabelled datasets with self-supervised learning (SSL), but this introduces the issue of domain shift if the pretraining and task dataset distributions differ. Data augmentations are essential for improving the generalizability of SSL-pretrained models, but they tend to be either handcrafted or randomly applied. We use an adversarial model to generate masks as augmentations for 12-lead electrocardiogram (ECG) data, where masks learn to occlude diagnostically-relevant regions. Compared to random augmentations, models pretrained with adversarial masking reaches better accuracy under a domain shift condition and in data-scarce regimes on two diverse downstream tasks, arrhythmia classification and patient age estimation. Adversarial masking is competitive with and even reaches further improvements when combined with state-of-art ECG augmentation methods, 3KG and random lead masking (RLM), demonstrating the generalizability of our method.
0 Replies

Loading