TY - GEN
T1 - Knowledge Distillation for Semi-supervised Domain Adaptation
AU - Orbes-Arteaga, Mauricio
AU - Cardoso, Jorge
AU - Sørensen, Lauge
AU - Igel, Christian
AU - Ourselin, Sebastien
AU - Modat, Marc
AU - Nielsen, Mads
AU - Pai, Akshay
PY - 2019/1/1
Y1 - 2019/1/1
N2 - In the absence of sufficient data variation (e.g., scanner and protocol variability) in annotated data, deep neural networks (DNNs) tend to overfit during training. As a result, their performance is significantly lower on data from unseen sources compared to the performance on data from the same source as the training data. Semi-supervised domain adaptation methods can alleviate this problem by tuning networks to new target domains without the need for annotated data from these domains. Adversarial domain adaptation (ADA) methods are a popular choice that aim to train networks in such a way that the features generated are domain agnostic. However, these methods require careful dataset-specific selection of hyperparameters such as the complexity of the discriminator in order to achieve a reasonable performance. We propose to use knowledge distillation (KD) – an efficient way of transferring knowledge between different DNNs – for semi-supervised domain adaption of DNNs. It does not require dataset-specific hyperparameter tuning, making it generally applicable. The proposed method is compared to ADA for segmentation of white matter hyperintensities (WMH) in magnetic resonance imaging (MRI) scans generated by scanners that are not a part of the training set. Compared with both the baseline DNN (trained on source domain only and without any adaption to target domain) and with using ADA for semi-supervised domain adaptation, the proposed method achieves significantly higher WMH dice scores.
AB - In the absence of sufficient data variation (e.g., scanner and protocol variability) in annotated data, deep neural networks (DNNs) tend to overfit during training. As a result, their performance is significantly lower on data from unseen sources compared to the performance on data from the same source as the training data. Semi-supervised domain adaptation methods can alleviate this problem by tuning networks to new target domains without the need for annotated data from these domains. Adversarial domain adaptation (ADA) methods are a popular choice that aim to train networks in such a way that the features generated are domain agnostic. However, these methods require careful dataset-specific selection of hyperparameters such as the complexity of the discriminator in order to achieve a reasonable performance. We propose to use knowledge distillation (KD) – an efficient way of transferring knowledge between different DNNs – for semi-supervised domain adaption of DNNs. It does not require dataset-specific hyperparameter tuning, making it generally applicable. The proposed method is compared to ADA for segmentation of white matter hyperintensities (WMH) in magnetic resonance imaging (MRI) scans generated by scanners that are not a part of the training set. Compared with both the baseline DNN (trained on source domain only and without any adaption to target domain) and with using ADA for semi-supervised domain adaptation, the proposed method achieves significantly higher WMH dice scores.
KW - Domain adaptation
KW - Knowledge distillation
KW - Semi-supervised learning
KW - White matter hyperintensities
UR - http://www.scopus.com/inward/record.url?scp=85075551780&partnerID=8YFLogxK
U2 - 10.1007/978-3-030-32695-1_8
DO - 10.1007/978-3-030-32695-1_8
M3 - Article in proceedings
AN - SCOPUS:85075551780
SN - 9783030326944
T3 - Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
SP - 68
EP - 76
BT - OR 2.0 Context-Aware Operating Theaters and Machine Learning in Clinical Neuroimaging - 2nd International Workshop, OR 2.0 2019, and 2nd International Workshop, MLCN 2019, Held in Conjunction with MICCAI 2019, Proceedings
A2 - Zhou, Luping
A2 - Sarikaya, Duygu
A2 - Kia, Seyed Mostafa
A2 - Speidel, Stefanie
A2 - Malpani, Anand
A2 - Hashimoto, Daniel
A2 - Habes, Mohamad
A2 - Löfstedt, Tommy
A2 - Ritter, Kerstin
A2 - Wang, Hongzhi
PB - Springer VS
T2 - 2nd International Workshop on Context-Aware Surgical Theaters, OR 2.0 2019, and the 2nd International Workshop on Machine Learning in Clinical Neuroimaging, MLCN 2019, held in conjunction with the 22nd International Conference on Medical Image Computing and Computer Assisted Intervention, MICCAI 2019
Y2 - 17 October 2019 through 17 October 2019
ER -