51,346 research outputs found

    A deep learning integrated Lee-Carter model

    Get PDF
    In the field of mortality, the Lee–Carter based approach can be considered the milestone to forecast mortality rates among stochastic models. We could define a “Lee–Carter model family” that embraces all developments of this model, including its first formulation (1992) that remains the benchmark for comparing the performance of future models. In the Lee–Carter model, the kt parameter, describing the mortality trend over time, plays an important role about the future mortality behavior. The traditional ARIMA process usually used to model kt shows evident limitations to describe the future mortality shape. Concerning forecasting phase, academics should approach a more plausible way in order to think a nonlinear shape of the projected mortality rates. Therefore, we propose an alternative approach the ARIMA processes based on a deep learning technique. More precisely, in order to catch the pattern of kt series over time more accurately, we apply a Recurrent Neural Network with a Long Short-Term Memory architecture and integrate the Lee–Carter model to improve its predictive capacity. The proposed approach provides significant performance in terms of predictive accuracy and also allow for avoiding the time-chunks’ a priori selection. Indeed, it is a common practice among academics to delete the time in which the noise is overflowing or the data quality is insufficient. The strength of the Long Short-Term Memory network lies in its ability to treat this noise and adequately reproduce it into the forecasted trend, due to its own architecture enabling to take into account significant long-term patterns

    Data-driven modeling of the olfactory neural codes and their dynamics in the insect antennal lobe

    Get PDF
    Recordings from neurons in the insects' olfactory primary processing center, the antennal lobe (AL), reveal that the AL is able to process the input from chemical receptors into distinct neural activity patterns, called olfactory neural codes. These exciting results show the importance of neural codes and their relation to perception. The next challenge is to \emph{model the dynamics} of neural codes. In our study, we perform multichannel recordings from the projection neurons in the AL driven by different odorants. We then derive a neural network from the electrophysiological data. The network consists of lateral-inhibitory neurons and excitatory neurons, and is capable of producing unique olfactory neural codes for the tested odorants. Specifically, we (i) design a projection, an odor space, for the neural recording from the AL, which discriminates between distinct odorants trajectories (ii) characterize scent recognition, i.e., decision-making based on olfactory signals and (iii) infer the wiring of the neural circuit, the connectome of the AL. We show that the constructed model is consistent with biological observations, such as contrast enhancement and robustness to noise. The study answers a key biological question in identifying how lateral inhibitory neurons can be wired to excitatory neurons to permit robust activity patterns

    Quantum Generative Adversarial Networks for Learning and Loading Random Distributions

    Full text link
    Quantum algorithms have the potential to outperform their classical counterparts in a variety of tasks. The realization of the advantage often requires the ability to load classical data efficiently into quantum states. However, the best known methods require O(2n)\mathcal{O}\left(2^n\right) gates to load an exact representation of a generic data structure into an nn-qubit state. This scaling can easily predominate the complexity of a quantum algorithm and, thereby, impair potential quantum advantage. Our work presents a hybrid quantum-classical algorithm for efficient, approximate quantum state loading. More precisely, we use quantum Generative Adversarial Networks (qGANs) to facilitate efficient learning and loading of generic probability distributions -- implicitly given by data samples -- into quantum states. Through the interplay of a quantum channel, such as a variational quantum circuit, and a classical neural network, the qGAN can learn a representation of the probability distribution underlying the data samples and load it into a quantum state. The loading requires O(poly(n))\mathcal{O}\left(poly\left(n\right)\right) gates and can, thus, enable the use of potentially advantageous quantum algorithms, such as Quantum Amplitude Estimation. We implement the qGAN distribution learning and loading method with Qiskit and test it using a quantum simulation as well as actual quantum processors provided by the IBM Q Experience. Furthermore, we employ quantum simulation to demonstrate the use of the trained quantum channel in a quantum finance application.Comment: 14 pages, 13 figure

    Cortical Networks for Control of Voluntary Arm Movements Under Variable Force Conditions

    Full text link
    A neural model of voluntary movement and proprioception functionally interprets and simulates cell types in movement related areas of primate cortex. The model circuit maintains accurate proprioception while controlling voluntary reaches to spatial targets, exertion of force against obstacles, posture maintenance despite perturbations, compliance with an imposed movement, and static and inertial load compensations. Computer simulations show that model cell properties mimic cell properties in areas 4 and 5. These include delay period activation, response profiles during movement, kinematic and kinetic sensitivities, and latency of activity onset. Model area 4 phasic and tonic cells compute velocity and position commands which activate alpha and gamma motor neurons, thereby shifting the mechanical equilibrium point. Anterior area 5 cells compute limb position using corollary discharges from area 4 and muscle spindle feedback. Posterior area 5 cells use the perceived position and target position signals to compute a desired movement vector. The cortical loop is closed by a volition-gated projection of this movement vector to area 4 phasic cells. Phasic-tonic cells in area 4 incorporate force command components to compensate for static and inertial loads. Predictions are made for both motor and parietal cell types under novel experimental protocols.Office of Naval Research (N00014-92-J-1309, N00014-93-1-1364, N00014-95-l-0409, N00014-92-J-4015); National Science Foundation (IRI-90-24877, IRI-90-00530
    • …
    corecore