TY - GEN
T1 - Semi-supervised learning via compact latent space clustering
AU - Kamnitsas, Konstantinos
AU - Castro, Daniel C.
AU - Le Folgoc, Loic
AU - Walker, Ian
AU - Tanno, Ryutaro
AU - Rueckert, Daniel
AU - Glocker, Ben
AU - Criminisi, Antonio
AU - Nori, Aditya
N1 - Publisher Copyright:
© CURRAN-CONFERENCE. All rights reserved.
PY - 2018/1/1
Y1 - 2018/1/1
N2 - We present a novel cost function for semisupervised learning of neural networks that encourages compact clustering of the latent space to facilitate separation. The key idea is to dynamically create a graph over embeddings of labeled and unlabeled samples of a training batch to capture underlying structure in feature space, and use label propagation to estimate its high and low density regions. We then devise a cost function based on Markov chains on the graph that regularizes the latent space to form a single compact cluster per class, while avoiding to disturb existing clusters during optimization. We evaluate our approach on three benchmarks and compare to state-of-the art with promising results. Our approach combines the benefits of graph-based regularization with efficient, inductive inference, does not require modifications to a network architecture, and can thus be easily applied to existing networks to enable an effective use of unlabeled data.
AB - We present a novel cost function for semisupervised learning of neural networks that encourages compact clustering of the latent space to facilitate separation. The key idea is to dynamically create a graph over embeddings of labeled and unlabeled samples of a training batch to capture underlying structure in feature space, and use label propagation to estimate its high and low density regions. We then devise a cost function based on Markov chains on the graph that regularizes the latent space to form a single compact cluster per class, while avoiding to disturb existing clusters during optimization. We evaluate our approach on three benchmarks and compare to state-of-the art with promising results. Our approach combines the benefits of graph-based regularization with efficient, inductive inference, does not require modifications to a network architecture, and can thus be easily applied to existing networks to enable an effective use of unlabeled data.
UR - https://www.scopus.com/pages/publications/85057226036
M3 - Conference contribution
AN - SCOPUS:85057226036
T3 - 35th International Conference on Machine Learning, ICML 2018
SP - 3845
EP - 3854
BT - 35th International Conference on Machine Learning, ICML 2018
A2 - Dy, Jennifer
A2 - Krause, Andreas
PB - International Machine Learning Society (IMLS)
T2 - 35th International Conference on Machine Learning, ICML 2018
Y2 - 10 July 2018 through 15 July 2018
ER -