248 research outputs found

    Toward the biological model of the hippocampus as the successor representation agent

    Full text link
    The hippocampus is an essential brain region for spatial memory and learning. Recently, a theoretical model of the hippocampus based on temporal difference (TD) learning has been published. Inspired by the successor representation (SR) learning algorithms, which decompose value function of TD learning into reward and state transition, they argued that the rate of firing of CA1 place cells in the hippocampus represents the probability of state transition. This theory, called predictive map theory, claims that the hippocampus representing space learns the probability of transition from the current state to the future state. The neural correlates of expecting the future state are the firing rates of the CA1 place cells. This explanation is plausible for the results recorded in behavioral experiments, but it is lacking the neurobiological implications. Modifying the SR learning algorithm added biological implications to the predictive map theory. Similar with the simultaneous needs of information of the current and future state in the SR learning algorithm, the CA1 place cells receive two inputs from CA3 and entorhinal cortex. Mathematical transformation showed that the SR learning algorithm is equivalent to the heterosynaptic plasticity rule. The heterosynaptic plasticity phenomena in CA1 were discussed and compared with the modified SR update rule. This study attempted to interpret the TD algorithm as the neurobiological mechanism occurring in place learning, and to integrate the neuroscience and artificial intelligence approaches in the field.Comment: 9 pages, 1 figur

    The effect of synaptic weight initialization in feature-based successor representation learning

    Full text link
    After discovering place cells, the idea of the hippocampal (HPC) function to represent geometric spaces has been extended to predictions, imaginations, and conceptual cognitive maps. Recent research arguing that the HPC represents a predictive map; and it has shown that the HPC predicts visits to specific locations. This predictive map theory is based on successor representation (SR) from reinforcement learning. Feature-based SR (SF), which uses a neural network as a function approximation to learn SR, seems more plausible neurobiological model. However, it is not well known how different methods of weight (W) initialization affect SF learning. In this study, SF learners were exposed to simple maze environments to analyze SF learning efficiency and W patterns pattern changes. Three kinds of W initialization pattern were used: identity matrix, zero matrix, and small random matrix. The SF learner initiated with random weight matrix showed better performance than other three RL agents. We will discuss the neurobiological meaning of SF weight matrix. Through this approach, this paper tried to increase our understanding of intelligence from neuroscientific and artificial intelligence perspective.Comment: 11 pages, 8 figures including 2 supplementary figure

    Insights from Analysis of Video Streaming Data to Improve Resource Management

    Full text link
    Today a large portion of Internet traffic is video. Over The Top (OTT) service providers offer video streaming services by creating a large distributed cloud network on top of a physical infrastructure owned by multiple entities. Our study explores insights from video streaming activity by analyzing data collected from Korea's largest OTT service provider. Our analysis of nationwide data shows interesting characteristics of video streaming such as correlation between user profile information (e.g., age, sex) and viewing habits, viewing habits of users (when do the users watch? using which devices?), viewing patterns (early leaving viewer vs. steady viewer), etc. Video on Demand (VoD) streaming involves costly (and often limited) compute, storage, and network resources. Findings from our study will be beneficial for OTTs, Content Delivery Networks (CDNs), Internet Service Providers (ISPs), and Carrier Network Operators, to improve their resource allocation and management techniques.Comment: This is a preprint electronic version of the article accepted to IEEE CloudNet 201

    Fabrication and Evaluation of Magnetic Micro actuators for Implantable Self-Clearing Glaucoma Drainage Devices

    Get PDF
    According to the World Health Organization, glaucoma is the second leading cause of blindness in the world. It currently affects more than 2.7 million people in the United States alone and over 79.6 million people worldwide are estimated to be inflicted by this debilitating disease by 2020. Glaucoma patients are often characterized with elevated intraocular pressure (IOP) and are treated with implantation of glaucoma drainage devices (GDD) to maintain optimum IOP. Although initially effective at delaying glaucoma progression, contemporary GDD often lead to numerous complications and only 50% of implanted devices remain functional after 5 years. Biofouling is seen to be one of the leading cause for the failure of GDDs. In order to overcome biological blockage, we propose a self-clearing glaucoma drainage device using integrated magnetic microactuators. Here we report on the maskless photolithographic fabrication results of magnetic microactuators to be integrated into a bespoke GDD. The maskless photography enabled rapid prototyping of microdevices using low-cost materials in contrast to conventional lithographic methods. The fabricated devices were able to produce maximum deflection of 57.9 degree at magnetic field strength of 32.6 kA/m. The static response of the fabricated devices was compared with the theoretical data

    Regularizing Towards Soft Equivariance Under Mixed Symmetries

    Full text link
    Datasets often have their intrinsic symmetries, and particular deep-learning models called equivariant or invariant models have been developed to exploit these symmetries. However, if some or all of these symmetries are only approximate, which frequently happens in practice, these models may be suboptimal due to the architectural restrictions imposed on them. We tackle this issue of approximate symmetries in a setup where symmetries are mixed, i.e., they are symmetries of not single but multiple different types and the degree of approximation varies across these types. Instead of proposing a new architectural restriction as in most of the previous approaches, we present a regularizer-based method for building a model for a dataset with mixed approximate symmetries. The key component of our method is what we call equivariance regularizer for a given type of symmetries, which measures how much a model is equivariant with respect to the symmetries of the type. Our method is trained with these regularizers, one per each symmetry type, and the strength of the regularizers is automatically tuned during training, leading to the discovery of the approximation levels of some candidate symmetry types without explicit supervision. Using synthetic function approximation and motion forecasting tasks, we demonstrate that our method achieves better accuracy than prior approaches while discovering the approximate symmetry levels correctly.Comment: Proceedings of the International Conference on Machine Learning (ICML), 202

    Intense ultraviolet emission from needle-like WO3 nanostructures synthesized by noncatalytic thermal evaporation

    Get PDF
    Photoluminescence measurements showed that needle-like tungsten oxide nanostructures synthesized at 590°C to 750°C by the thermal evaporation of WO3 nanopowders without the use of a catalyst had an intense near-ultraviolet (NUV) emission band that was different from that of the tungsten oxide nanostructures obtained in other temperature ranges. The intense NUV emission might be due to the localized states associated with oxygen vacancies and surface states

    Probabilistic Imputation for Time-series Classification with Missing Data

    Full text link
    Multivariate time series data for real-world applications typically contain a significant amount of missing values. The dominant approach for classification with such missing values is to impute them heuristically with specific values (zero, mean, values of adjacent time-steps) or learnable parameters. However, these simple strategies do not take the data generative process into account, and more importantly, do not effectively capture the uncertainty in prediction due to the multiple possibilities for the missing values. In this paper, we propose a novel probabilistic framework for classification with multivariate time series data with missing values. Our model consists of two parts; a deep generative model for missing value imputation and a classifier. Extending the existing deep generative models to better capture structures of time-series data, our deep generative model part is trained to impute the missing values in multiple plausible ways, effectively modeling the uncertainty of the imputation. The classifier part takes the time series data along with the imputed missing values and classifies signals, and is trained to capture the predictive uncertainty due to the multiple possibilities of imputations. Importantly, we show that na\"ively combining the generative model and the classifier could result in trivial solutions where the generative model does not produce meaningful imputations. To resolve this, we present a novel regularization technique that can promote the model to produce useful imputation values that help classification. Through extensive experiments on real-world time series data with missing values, we demonstrate the effectiveness of our method
    • …
    corecore