TY - JOUR
T1 - Explicitly disentangling image content from translation and rotation with spatial-VAE
AU - Bepler, Tristan
AU - Zhong, Ellen D.
AU - Kelley, Kotaro
AU - Brignole, Edward
AU - Berger, Bonnie
N1 - Funding Information:
This work was supported by NIH R01-GM081871. We would like to thank Bridget Carragher and Clint Potter at NYSBC for their support in providing the antibody dataset. The NYSBC portion of this work was supported by Simons Foundation SF349247, NYSTAR, and NIH GM103310 with additional support from Agouron Institute F00316 and NIH OD019994-01. We would also like to thank the laboratory of HHMI investigator Catherine L. Drennan, MIT, for providing the CODH/ACS dataset that was collected with support from the National Institutes of Health (R35 GM126982).
Funding Information:
This work was supported by NIH R01-GM081871.
Funding Information:
We would like to thank Bridget Carragher and Clint Potter at NYSBC for their support in providing the antibody dataset. The NYSBC portion of this work was supported by Simons Foundation SF349247, NYSTAR, and NIH GM103310 with additional support from Agouron Institute F00316 and NIH OD019994-01.
Publisher Copyright:
© 2019 Neural information processing systems foundation. All rights reserved.
PY - 2019
Y1 - 2019
N2 - Given an image dataset, we are often interested in finding data generative factors that encode semantic content independently from pose variables such as rotation and translation. However, current disentanglement approaches do not impose any specific structure on the learned latent representations. We propose a method for explicitly disentangling image rotation and translation from other unstructured latent factors in a variational autoencoder (VAE) framework. By formulating the generative model as a function of the spatial coordinate, we make the reconstruction error differentiable with respect to latent translation and rotation parameters. This formulation allows us to train a neural network to perform approximate inference on these latent variables while explicitly constraining them to only represent rotation and translation. We demonstrate that this framework, termed spatial-VAE, effectively learns latent representations that disentangle image rotation and translation from content and improves reconstruction over standard VAEs on several benchmark datasets, including applications to modeling continuous 2-D views of proteins from single particle electron microscopy and galaxies in astronomical images.
AB - Given an image dataset, we are often interested in finding data generative factors that encode semantic content independently from pose variables such as rotation and translation. However, current disentanglement approaches do not impose any specific structure on the learned latent representations. We propose a method for explicitly disentangling image rotation and translation from other unstructured latent factors in a variational autoencoder (VAE) framework. By formulating the generative model as a function of the spatial coordinate, we make the reconstruction error differentiable with respect to latent translation and rotation parameters. This formulation allows us to train a neural network to perform approximate inference on these latent variables while explicitly constraining them to only represent rotation and translation. We demonstrate that this framework, termed spatial-VAE, effectively learns latent representations that disentangle image rotation and translation from content and improves reconstruction over standard VAEs on several benchmark datasets, including applications to modeling continuous 2-D views of proteins from single particle electron microscopy and galaxies in astronomical images.
UR - http://www.scopus.com/inward/record.url?scp=85090175253&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85090175253&partnerID=8YFLogxK
M3 - Conference article
AN - SCOPUS:85090175253
SN - 1049-5258
VL - 32
JO - Advances in Neural Information Processing Systems
JF - Advances in Neural Information Processing Systems
T2 - 33rd Annual Conference on Neural Information Processing Systems, NeurIPS 2019
Y2 - 8 December 2019 through 14 December 2019
ER -