Variational Autoencoders For Semi-Supervised Deep Metric Learning
Nathan Safir
Meekail Zain
Curtis Godwin
Eric Miller
Bella Humphrey
Shannon P Quinn
Deep metric learning (DML) methods generally do not incorporate unlabelled data. We propose
borrowing components of the variational autoencoder (VAE) methodology to extend DML
methods to train on semi-supervised datasets. We experimentally evaluate the atomic benefits to the perform-
ing DML on the VAE latent space such as the enhanced ability to train using unlabelled data and to induce
bias given prior knowledge. We find that jointly training DML with an autoencoder and VAE may be potentially
helpful for some semi-suprevised datasets, but that a training routine of alternating between the DML loss
and an additional unsupervised loss across epochs is generally unviable.
Variational Autoencoders, Metric Learning, Deep Learning, Representation Learning, Generative Models
DOI10.25080/majora-212e5952-022