Skip to yearly menu bar Skip to main content


Poster

A Framework for the Quantitative Evaluation of Disentangled Representations

Cian Eastwood · Chris Williams

East Meeting level; 1,2,3 #16

Abstract:

Recent AI research has emphasised the importance of learning disentangled representations of the explanatory factors behind data. Despite the growing interest in models which can learn such representations, visual inspection remains the standard evaluation metric. While various desiderata have been implied in recent definitions, it is currently unclear what exactly makes one disentangled representation better than another. In this work we propose a framework for the quantitative evaluation of disentangled representations when the ground-truth latent structure is available. Three criteria are explicitly defined and quantified to elucidate the quality of learnt representations and thus compare models on an equal basis. To illustrate the appropriateness of the framework, we employ it to compare quantitatively the representations learned by recent state-of-the-art models.

Live content is unavailable. Log in and register to view live content