Learned Equivariant Rendering without Transformation Supervision

11/11/2020
by   Cinjon Resnick, et al.
4

We propose a self-supervised framework to learn scene representations from video that are automatically delineated into objects and background. Our method relies on moving objects being equivariant with respect to their transformation across frames and the background being constant. After training, we can manipulate and render the scenes in real time to create unseen combinations of objects, transformations, and backgrounds. We show results on moving MNIST with backgrounds.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset