Neuromorphic computing leverages the complex dynamics of physical systems for
computation. The field has recently undergone an explosion in the range and
sophistication of implementations, with rapidly improving performance.
Neuromorphic schemes typically employ a single physical system, limiting the
dimensionality and range of available dynamics - restricting strong performance
to a few specific tasks. This is a critical roadblock facing the field,
inhibiting the power and versatility of neuromorphic schemes.
Here, we present a solution. We engineer a diverse suite of nanomagnetic
arrays and show how tuning microstate space and geometry enables a broad range
of dynamics and computing performance. We interconnect arrays in parallel,
series and multilayered neural network architectures, where each network node
is a distinct physical system. This networked approach grants extremely high
dimensionality and enriched dynamics enabling meta-learning to be implemented
on small training sets and exhibiting strong performance across a broad
taskset. We showcase network performance via few-shot learning, rapidly
adapting on-the-fly to previously unseen tasks