Overview
Many physical, biological and engineering systems evolve over time according to geometric laws, for example planets follow elliptical orbits shaped by gravity, and fluids swirl along streamlines governed by curvature and vorticity. A central challenge in modern machine learning is to model these dynamics faithfully and efficiently, especially in cases where the underlying structure is nonlinear, high-dimensional, or only indirectly observed.
Most existing deep learning approaches to dynamics approximate trajectories in flat, Euclidean spaces and often overlook the underlying geometry of the data. This leads to models that may be expressive but lack interpretability, physical consistency, or generalisation capacity.
In this preliminary project, we introduce neural geodesic flows – a new class of machine learning models that learn to predict dynamical systems by assuming the system evolves along the geodesics of a learned latent Riemannian manifold. This approach constrains the model with structure-preserving geometric properties that adapt naturally to complex dynamics.
Research highlights
Neural geodesic flows
To predict the dynamics of a system, we first map observations (such as the positions of two planets, or the velocity of a fluid) to coordinates on a latent manifold. Next, we evolve the coordinates over time by moving the system along geodesics (locally length-minimising curves) on the manifold. This is done by solving the geodesic flow equations on the manifold. Finally, we map the manifold coordinates back to the observation (data) space to predict the dynamics.
Importantly, we jointly learn the mapping back and forth between observations and coordinates on the latent manifold, as well as the metric of the manifold, by representing them with deep neural networks and training the model to match example trajectories of the system in the data space. An example of our model learning the toy dynamics of particles flowing over the surface of a sphere is shown in Fig. 1.

We call this model a neural geodesic flow and its schematic is shown in Fig. 2.

Our core assumption is that the observed data evolves along geodesics of a latent manifold. This may seem strong, but multiple physical systems can be interpreted in this way, including classical mechanics and spacetime evolution in general relativity, with deep links to the principle of least action. By construction our model naturally conserves geodesic energy, which, when trained, can lead to good total energy conservation. An example of our model learning to model the motion of two bodies orbiting each other is shown in Fig. 3.

Applications and impact
Neural geodesic flows are designed to model the intrinsic geometrical structure of complex dynamical data. Future applications could include learning latent physics models, compressing simulations to low-dimensional embeddings, and more generally advancing manifold learning models.
Future directions
This is a preliminary project and we actively working on improving this model. Please contact us if you are interested in contributing / learning more about this project.
Publications
Julian Bürge. Neural Geodesic Flows. Master’s Thesis, ETH Zurich (2025).
Resources
Team & collaborators
- Julian Bürge: ETH Zurich.
- Siddhartha Mishra: ETH Zurich.
- Ben Moseley: Imperial College London.
- Lewis O’Donnell: Imperial College London.