53 research outputs found
Scale-invariant temporal history (SITH): optimal slicing of the past in an uncertain world
In both the human brain and any general artificial intelligence (AI), a
representation of the past is necessary to predict the future. However, perfect
storage of all experiences is not possible. One possibility, utilized in many
applications, is to retain information about the past in a buffer. A limitation
of this approach is that although events in the buffer are represented with
perfect accuracy, the resources necessary to represent information at a
particular time scale go up rapidly. Here we present a neurally-plausible,
compressed, scale-free memory representation we call Scale-Invariant Temporal
History (SITH). This representation covers an exponentially large period of
time in the past at the cost of sacrificing temporal accuracy for events
further in the past. The form of this decay is scale-invariant and can be shown
to be optimal in that it is able to respond to worlds with a wide range of time
scales. We demonstrate the utility of this representation in learning to play a
simple video game. In this environment, SITH exhibits better learning
performance than a fixed-size buffer history representation. Whereas the buffer
performs well as long as the temporal dependencies can be represented within
the buffer, SITH performs well over a much larger range of time scales for the
same amount of resources. Finally, we discuss how the application of SITH,
along with other human-inspired models of cognition, could improve
reinforcement and machine learning algorithms in general.First author draf
DeepSITH: efficient learning via decomposition of what and when across time scales
Extracting temporal relationships over a range of scales is a hallmark of human perception and cognition -- and thus it is a critical feature of machine learning applied to real-world problems. Neural networks are either plagued by the exploding/vanishing gradient problem in recurrent neural networks (RNNs) or must adjust their parameters to learn the relevant time scales (e.g., in LSTMs). This paper introduces DeepSITH, a network comprising biologically-inspired Scale-Invariant Temporal History (SITH) modules in series with dense connections between layers. SITH modules respond to their inputs with a geometrically-spaced set of time constants, enabling the DeepSITH network to learn problems along a continuum of time-scales. We compare DeepSITH to LSTMs and other recent RNNs on several time series prediction and decoding tasks. DeepSITH achieves state-of-the-art performance on these problems.https://papers.nips.cc/paper/2021/file/e7dfca01f394755c11f853602cb2608a-Paper.pdfPublished versio
PyMVPA: A Unifying Approach to the Analysis of Neuroscientific Data
The Python programming language is steadily increasing in popularity as the language of choice for scientific computing. The ability of this scripting environment to access a huge code base in various languages, combined with its syntactical simplicity, make it the ideal tool for implementing and sharing ideas among scientists from numerous fields and with heterogeneous methodological backgrounds. The recent rise of reciprocal interest between the machine learning (ML) and neuroscience communities is an example of the desire for an inter-disciplinary transfer of computational methods that can benefit from a Python-based framework. For many years, a large fraction of both research communities have addressed, almost independently, very high-dimensional problems with almost completely non-overlapping methods. However, a number of recently published studies that applied ML methods to neuroscience research questions attracted a lot of attention from researchers from both fields, as well as the general public, and showed that this approach can provide novel and fruitful insights into the functioning of the brain. In this article we show how PyMVPA, a specialized Python framework for machine learning based data analysis, can help to facilitate this inter-disciplinary technology transfer by providing a single interface to a wide array of machine learning libraries and neural data-processing methods. We demonstrate the general applicability and power of PyMVPA via analyses of a number of neural data modalities, including fMRI, EEG, MEG, and extracellular recordings
- …