When neural networks are trained from data to simulate the dynamics of
physical systems, they encounter a persistent challenge: the long-time dynamics
they produce are often unphysical or unstable. We analyze the origin of such
instabilities when learning linear dynamical systems, focusing on the training
dynamics. We make several analytical findings which empirical observations
suggest extend to nonlinear dynamical systems. First, the rate of convergence
of the training dynamics is uneven and depends on the distribution of energy in
the data. As a special case, the dynamics in directions where the data have no
energy cannot be learned. Second, in the unlearnable directions, the dynamics
produced by the neural network depend on the weight initialization, and common
weight initialization schemes can produce unstable dynamics. Third, injecting
synthetic noise into the data during training adds damping to the training
dynamics and can stabilize the learned simulator, though doing so undesirably
biases the learned dynamics. For each contributor to instability, we suggest
mitigative strategies. We also highlight important differences between learning
discrete-time and continuous-time dynamics, and discuss extensions to nonlinear
systems.Comment: 15 page