Skip to yearly menu bar Skip to main content


In-Person Poster presentation / poster accept

Latent Graph Inference using Product Manifolds

Haitz Sáez de Ocáriz Borde · Anees Kazi · Federico Barbero · Pietro Lio

MH1-2-3-4 #50

Keywords: [ Deep Learning and representational learning ] [ Product Manifolds ] [ Latent Graph Inference ] [ graph neural networks ]


Abstract:

Graph Neural Networks usually rely on the assumption that the graph topology is available to the network as well as optimal for the downstream task. Latent graph inference allows models to dynamically learn the intrinsic graph structure of problems where the connectivity patterns of data may not be directly accessible. In this work, we generalize the discrete Differentiable Graph Module (dDGM) for latent graph learning. The original dDGM architecture used the Euclidean plane to encode latent features based on which the latent graphs were generated. By incorporating Riemannian geometry into the model and generating more complex embedding spaces, we can improve the performance of the latent graph inference system. In particular, we propose a computationally tractable approach to produce product manifolds of constant curvature model spaces that can encode latent features of varying structure. The latent representations mapped onto the inferred product manifold are used to compute richer similarity measures that are leveraged by the latent graph learning model to obtain optimized latent graphs. Moreover, the curvature of the product manifold is learned during training alongside the rest of the network parameters and based on the downstream task, rather than it being a static embedding space. Our novel approach is tested on a wide range of datasets, and outperforms the original dDGM model.

Chat is not available.