Variational Integrator Networks for Physically Structured Embeddings

Abstract

Learning workable representations of dynamical systems is becoming an increasingly important problem in a number of application areas. By leveraging recent work connecting deep neural networks to systems of differential equations, we propose variational integrator networks, a class of neural network architectures designed to preserve the geometric structure of physical systems. This class of network architectures facilitates accurate long-term prediction, interpretability, and data-efficient learning, while still remaining highly flexible and capable of modeling complex behavior. We demonstrate that they both noisy observations in phase space and from image pixels within which the unknown dynamics are embedded.

Publication
Proceedings of the International Conference on Artificial Intelligence and Statistics (AISTATS)
Steindór Sæmundsson
Steindór Sæmundsson
PhD (11/2016-11/2021)
Alexander Terenin
Alexander Terenin
PhD (10/2018-11/2021)
Marc Deisenroth
Marc Deisenroth
Google DeepMind Chair of Machine Learning and Artificial Intelligence