Evaluation of metric and representation learning approaches: Effects of representations driven by relative distance on the performance.

Researchers

Journal

Modalities

Models

Abstract

Several deep neural network architectures have emerged recently for metric learning. We asked which architecture is the most effective in measuring the similarity or dissimilarity among images. To this end, we evaluated six networks on a standard image set. We evaluated variational autoencoders, Siamese networks, triplet networks, and variational auto-encoders combined with Siamese or triplet networks. These networks were compared to a baseline network consisting of multiple separable convolutional layers. Our study revealed the following: (i) the triplet architecture proved the most effective one due to learning a relative distance – not an absolute distance; (ii) combining auto-encoders with networks that learn metrics (e.g., Siamese or triplet networks) is unwarranted; and (iii) an architecture based on separable convolutional layers is a reasonable simple alternative to triplet networks. These results can potentially impact our field by encouraging architects to develop advanced networks that take advantage of separable convolution and relative distance.

Similar Posts

Leave a Reply

Your email address will not be published. Required fields are marked *