TY - GEN
T1 - An Extensive Pixel-Level Augmentation Framework for Unsupervised Cross-Modality Domain Adaptation
AU - Baldeon Calisto, Maria G.
AU - Lai-Yuen, Susana K.
N1 - Publisher Copyright:
© 2023 SPIE.
PY - 2023/4/3
Y1 - 2023/4/3
N2 - Convolutional neural networks (CNNs) have achieved great success in automating the segmentation of medical images. Nevertheless, when a trained CNN is tested on a new domain there is a performance degradation due to the distribution shift. In this work, we present an unsupervised Extensive Pixel-level Augmentation framework (EPA) for cross-modality domain adaptation. EPA implements a two-phase image- and feature-level adaptation method. In the first phase, the source domain images are mapped to target domain in pixel space using the CycleGAN, StAC-DA, and CUT translation models. This creates an augmented translated dataset 3 times bigger than the original. In phase 2, a deeply supervised U-Net network is trained to segment the target images using a semi-supervised adversarial learning approach. In particular, a set of discriminator networks are trained to distinguish between the target and source domain segmentations, while the U-Net aims to fool them. EPA is tested on the task of brain structure segmentation from the Crossmoda 2022 Grand Challenge, being ranked within the top 12 submissions of the testing phase. Moreover, we demonstrate that augmenting the size of the mapped dataset through distinct translation methods is crucial for increasing the segmentation accuracy of the model.
AB - Convolutional neural networks (CNNs) have achieved great success in automating the segmentation of medical images. Nevertheless, when a trained CNN is tested on a new domain there is a performance degradation due to the distribution shift. In this work, we present an unsupervised Extensive Pixel-level Augmentation framework (EPA) for cross-modality domain adaptation. EPA implements a two-phase image- and feature-level adaptation method. In the first phase, the source domain images are mapped to target domain in pixel space using the CycleGAN, StAC-DA, and CUT translation models. This creates an augmented translated dataset 3 times bigger than the original. In phase 2, a deeply supervised U-Net network is trained to segment the target images using a semi-supervised adversarial learning approach. In particular, a set of discriminator networks are trained to distinguish between the target and source domain segmentations, while the U-Net aims to fool them. EPA is tested on the task of brain structure segmentation from the Crossmoda 2022 Grand Challenge, being ranked within the top 12 submissions of the testing phase. Moreover, we demonstrate that augmenting the size of the mapped dataset through distinct translation methods is crucial for increasing the segmentation accuracy of the model.
KW - Data Augmentation
KW - Domain Adaptation
KW - Generative Adversarial Networks
KW - Image Segmentation
KW - Medical Image Analysis
KW - Unsupervised Learning
UR - http://www.scopus.com/inward/record.url?scp=85159711131&partnerID=8YFLogxK
U2 - 10.1117/12.2651462
DO - 10.1117/12.2651462
M3 - Contribución a la conferencia
AN - SCOPUS:85159711131
T3 - Medical Imaging 2023: Image Processing
BT - Medical Imaging 2023
A2 - Colliot, Olivier
A2 - Isgum, Ivana
PB - SPIE
T2 - Medical Imaging 2023: Image Processing
Y2 - 19 February 2023 through 23 February 2023
ER -