Abstract
Differential equations parameterized by neural networks become expensive to
solve numerically as training progresses. We propose a remedy that encourages
learned dynamics to be easier to solve. Specifically, we introduce a
differentiable surrogate for the time cost of standard numerical solvers, using
higher-order derivatives of solution trajectories. These derivatives are
efficient to compute with Taylor-mode automatic differentiation. Optimizing
this additional objective trades model performance against the time cost of
solving the learned dynamics. We demonstrate our approach by training
substantially faster, while nearly as accurate, models in supervised
classification, density estimation, and time-series modelling tasks.
Users
Please
log in to take part in the discussion (add own reviews or comments).