MaxStyle: adversarial style composition for robust medical image segmentation
File(s)2206.01737v1.pdf (2.48 MB)
Accepted version
Author(s)
Type
Conference Paper
Abstract
Convolutional neural networks (CNNs) have achieved remarkable segmentation
accuracy on benchmark datasets where training and test sets are from the same
domain, yet their performance can degrade significantly on unseen domains,
which hinders the deployment of CNNs in many clinical scenarios. Most existing
works improve model out-of-domain (OOD) robustness by collecting multi-domain
datasets for training, which is expensive and may not always be feasible due to
privacy and logistical issues. In this work, we focus on improving model
robustness using a single-domain dataset only. We propose a novel data
augmentation framework called MaxStyle, which maximizes the effectiveness of
style augmentation for model OOD performance. It attaches an auxiliary
style-augmented image decoder to a segmentation network for robust feature
learning and data augmentation. Importantly, MaxStyle augments data with
improved image style diversity and hardness, by expanding the style space with
noise and searching for the worst-case style composition of latent features via
adversarial training. With extensive experiments on multiple public cardiac and
prostate MR datasets, we demonstrate that MaxStyle leads to significantly
improved out-of-distribution robustness against unseen corruptions as well as
common distribution shifts across multiple, different, unseen sites and unknown
image sequences under both low- and high-training data settings. The code can
be found at https://github.com/cherise215/MaxStyle.
accuracy on benchmark datasets where training and test sets are from the same
domain, yet their performance can degrade significantly on unseen domains,
which hinders the deployment of CNNs in many clinical scenarios. Most existing
works improve model out-of-domain (OOD) robustness by collecting multi-domain
datasets for training, which is expensive and may not always be feasible due to
privacy and logistical issues. In this work, we focus on improving model
robustness using a single-domain dataset only. We propose a novel data
augmentation framework called MaxStyle, which maximizes the effectiveness of
style augmentation for model OOD performance. It attaches an auxiliary
style-augmented image decoder to a segmentation network for robust feature
learning and data augmentation. Importantly, MaxStyle augments data with
improved image style diversity and hardness, by expanding the style space with
noise and searching for the worst-case style composition of latent features via
adversarial training. With extensive experiments on multiple public cardiac and
prostate MR datasets, we demonstrate that MaxStyle leads to significantly
improved out-of-distribution robustness against unseen corruptions as well as
common distribution shifts across multiple, different, unseen sites and unknown
image sequences under both low- and high-training data settings. The code can
be found at https://github.com/cherise215/MaxStyle.
Date Issued
2022-09-16
Date Acceptance
2022-05-05
Citation
Lecture Notes in Computer Science, 2022, 13435, pp.151-161
ISBN
978-3-031-16442-2
Publisher
Springer
Start Page
151
End Page
161
Journal / Book Title
Lecture Notes in Computer Science
Volume
13435
Copyright Statement
© 2022 The Author(s), under exclusive license to Springer Nature Switzerland AG.
Identifier
http://arxiv.org/abs/2206.01737v1
Source
Medical Image Computing and Computer Assisted Interventions (MICCAI) 2022
Subjects
cs.CV
eess.IV
eess.IV
q-bio.QM
Notes
Early accepted by MICCAI 2022
Publication Status
Published
Start Date
2022-09-18
Finish Date
2022-09-22
Coverage Spatial
Singapore