Pulling back information geometry

Georgios Arvanitidis, Miguel Gonzalez Duque, Alison Pouplin, Dimitris Kalatzis, Søren Hauberg

Publikation: Artikel i tidsskrift og konference artikel i tidsskriftKonferenceartikelForskningpeer review

Abstract

Latent space geometry has shown itself to provide a rich and rigorous framework for interacting with the latent variables of deep generative models. The existing theory, however, relies on the decoder being a Gaussian distribution as its simple reparametrization allows us to interpret the generating process as a random projection of a deterministic manifold. Consequently, this approach breaks down when applied to decoders that are not as easily reparametrized. We here propose to use the Fisher-Rao metric associated with the space of decoder distributions as a reference metric, which we pull back to the latent space. We show that we can achieve meaningful latent geometries for a wide range of decoder distributions for which the previous theory was not applicable, opening the door to ’black box’ latent geometries.
OriginalsprogEngelsk
TidsskriftProceedings of the 25th International Conference on Artificial Intelligence and Statistics (AISTATS) 2022
Vol/bind151
Sider (fra-til)4872
Antal sider4.894
StatusUdgivet - 2022

Fingeraftryk

Dyk ned i forskningsemnerne om 'Pulling back information geometry'. Sammen danner de et unikt fingeraftryk.

Citationsformater