Convolutional neural networks (CNN) are a powerful deep learning method for medical image segmentation. However performance in clinical practice often deteriorates when attempting to generalize models trained from a particular source domain to a different target domain (e.g. different vendor, acquisition parameters, protocols). To address this issue, domain adaptation has attracted increasing attention because it can minimize distribution differences among different but related domains. Extending from this prior work, we introduce Co-Unet-GAN, a co-learning domain adaptation and segmentation model addressing the domain shift problem. In this model, we train a Unet segmentation network and an image translation generative adversarial network (GAN) together to generalize performance across domains given supervised data only in the source domain. We evaluate our model on two large open echocardiography datasets, using the CAMUS set as supervised source domain and EchoNet-Dynamic as the unsupervised target. In this context, we obtain mean absolute error on ejection fraction of 9.67% on Co-Unet-GAN compared to 11.28% for a previously published Unet-GAN. Our Co-Unet-GAN for image translation and segmentation is a promising solution to the domain shift problem.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
INSTITUTIONAL Select your institution to access the SPIE Digital Library.
PERSONAL Sign in with your SPIE account to access your personal subscriptions or to use specific features such as save to my library, sign up for alerts, save searches, etc.