TY - JOUR
T1 - Semi-Supervised Learning via Compact Latent Space Clustering
AU - Kamnitsas, Konstantinos
AU - Castro, Daniel C.
AU - Folgoc, Loic Le
AU - Walker, Ian
AU - Tanno, Ryutaro
AU - Rueckert, Daniel
AU - Glocker, Ben
AU - Criminisi, Antonio
AU - Nori, Aditya
N1 - Publisher Copyright:
© 2018 by the author(s).
PY - 2018/1/1
Y1 - 2018/1/1
N2 - We present a novel cost function for semi-supervised learning of neural networks that encourages compact clustering of the latent space to facilitate separation. The key idea is to dynamically create a graph over embeddings of labeled and unlabeled samples of a training batch to capture underlying structure in feature space, and use label propagation to estimate its high and low density regions. We then devise a cost function based on Markov chains on the graph that regularizes the latent space to form a single compact cluster per class, while avoiding to disturb existing clusters during optimization. We evaluate our approach on three benchmarks and compare to state-of-the art with promising results. Our approach combines the benefits of graph-based regularization with efficient, inductive inference, does not require modifications to a network architecture, and can thus be easily applied to existing networks to enable an effective use of unlabeled data.
AB - We present a novel cost function for semi-supervised learning of neural networks that encourages compact clustering of the latent space to facilitate separation. The key idea is to dynamically create a graph over embeddings of labeled and unlabeled samples of a training batch to capture underlying structure in feature space, and use label propagation to estimate its high and low density regions. We then devise a cost function based on Markov chains on the graph that regularizes the latent space to form a single compact cluster per class, while avoiding to disturb existing clusters during optimization. We evaluate our approach on three benchmarks and compare to state-of-the art with promising results. Our approach combines the benefits of graph-based regularization with efficient, inductive inference, does not require modifications to a network architecture, and can thus be easily applied to existing networks to enable an effective use of unlabeled data.
UR - https://www.scopus.com/pages/publications/105020572070
M3 - Conference article
AN - SCOPUS:105020572070
SN - 2640-3498
VL - 80
SP - 2459
EP - 2468
JO - Proceedings of Machine Learning Research
JF - Proceedings of Machine Learning Research
T2 - 35th International Conference on Machine Learning, ICML 2018
Y2 - 10 July 2018 through 15 July 2018
ER -