272,380 research outputs found

    Improving M-Learners\u27 Performance through Deep Learning Techniques by Leveraging Features Weights

    Get PDF
    Β© 2013 IEEE. Mobile learning (M-learning) has gained tremendous attention in the educational environment in the past decade. For effective M-learning, it is important to create an efficient M-learning model that can identify the exact requirements of mobile learners (M-learners). M-learning model is composed of features that are generated during M-learners\u27 interaction with mobile devices. For an adaptive M-learning model, not only learning features are required, but it is also important to determine how they differ for various M-learners, their weights, and interrelationship. This study proposes a robust and adaptive M-learning model that is based on machine learning and deep learning (ML/DL) techniques. The proposed M-learning model dynamically explores learning features, their corresponding weights, and association for M-learners. Based on learning features, the M-learning model categorizes M-learners into different performance groups. The M-learning model then provides adaptive content, suggestions, and recommendations to M-learners in order to make learning adaptive and stimulating. For comparative analysis, the prediction accuracy of five baseline ML models was compared with the deep Artificial Neural Network (deep ANN). The results demonstrated that deep ANN and Random Forest (RF) models exhibited better prediction accuracy. Subsequently, both models were selected for developing the M-learning model which included the performance categorization of M-learners under a five-level classification scheme and assigning weights to various features for providing adaptive help and support to M-learners. Our explanatory analysis has shown that behavioral features besides contextual features also influence the learning performance of M-learners. As a direct outcome of this research, more efficient, interactive, and useful mobile learning applications can be developed that accurately predict learning objectives and requirements of diverse M-learners thus helping M-learners in enhancing their study behavior

    Learning to Learn Kernels with Variational Random Features

    Get PDF
    In this work, we introduce kernels with random Fourier features in the meta-learning framework to leverage their strong few-shot learning ability. We propose meta variational random features (MetaVRF) to learn adaptive kernels for the base-learner, which is developed in a latent variable model by treating the random feature basis as the latent variable. We formulate the optimization of MetaVRF as a variational inference problem by deriving an evidence lower bound under the meta-learning framework. To incorporate shared knowledge from related tasks, we propose a context inference of the posterior, which is established by an LSTM architecture. The LSTM-based inference network can effectively integrate the context information of previous tasks with task-specific information, generating informative and adaptive features. The learned MetaVRF can produce kernels of high representational power with a relatively low spectral sampling rate and also enables fast adaptation to new tasks. Experimental results on a variety of few-shot regression and classification tasks demonstrate that MetaVRF delivers much better, or at least competitive, performance compared to existing meta-learning alternatives.Comment: ICML'2020; code is available in: https://github.com/Yingjun-Du/MetaVR

    Intelligent Control of a Sensor-Actuator System via Kernelized Least-Squares Policy Iteration

    Get PDF
    In this paper a new framework, called Compressive Kernelized Reinforcement Learning (CKRL), for computing near-optimal policies in sequential decision making with uncertainty is proposed via incorporating the non-adaptive data-independent Random Projections and nonparametric Kernelized Least-squares Policy Iteration (KLSPI). Random Projections are a fast, non-adaptive dimensionality reduction framework in which high-dimensionality data is projected onto a random lower-dimension subspace via spherically random rotation and coordination sampling. KLSPI introduce kernel trick into the LSPI framework for Reinforcement Learning, often achieving faster convergence and providing automatic feature selection via various kernel sparsification approaches. In this approach, policies are computed in a low-dimensional subspace generated by projecting the high-dimensional features onto a set of random basis. We first show how Random Projections constitute an efficient sparsification technique and how our method often converges faster than regular LSPI, while at lower computational costs. Theoretical foundation underlying this approach is a fast approximation of Singular Value Decomposition (SVD). Finally, simulation results are exhibited on benchmark MDP domains, which confirm gains both in computation time and in performance in large feature spaces

    Improved Dropout for Shallow and Deep Learning

    Full text link
    Dropout has been witnessed with great success in training deep neural networks by independently zeroing out the outputs of neurons at random. It has also received a surge of interest for shallow learning, e.g., logistic regression. However, the independent sampling for dropout could be suboptimal for the sake of convergence. In this paper, we propose to use multinomial sampling for dropout, i.e., sampling features or neurons according to a multinomial distribution with different probabilities for different features/neurons. To exhibit the optimal dropout probabilities, we analyze the shallow learning with multinomial dropout and establish the risk bound for stochastic optimization. By minimizing a sampling dependent factor in the risk bound, we obtain a distribution-dependent dropout with sampling probabilities dependent on the second order statistics of the data distribution. To tackle the issue of evolving distribution of neurons in deep learning, we propose an efficient adaptive dropout (named \textbf{evolutional dropout}) that computes the sampling probabilities on-the-fly from a mini-batch of examples. Empirical studies on several benchmark datasets demonstrate that the proposed dropouts achieve not only much faster convergence and but also a smaller testing error than the standard dropout. For example, on the CIFAR-100 data, the evolutional dropout achieves relative improvements over 10\% on the prediction performance and over 50\% on the convergence speed compared to the standard dropout.Comment: In NIPS 201

    A Family of Maximum Margin Criterion for Adaptive Learning

    Full text link
    In recent years, pattern analysis plays an important role in data mining and recognition, and many variants have been proposed to handle complicated scenarios. In the literature, it has been quite familiar with high dimensionality of data samples, but either such characteristics or large data have become usual sense in real-world applications. In this work, an improved maximum margin criterion (MMC) method is introduced firstly. With the new definition of MMC, several variants of MMC, including random MMC, layered MMC, 2D^2 MMC, are designed to make adaptive learning applicable. Particularly, the MMC network is developed to learn deep features of images in light of simple deep networks. Experimental results on a diversity of data sets demonstrate the discriminant ability of proposed MMC methods are compenent to be adopted in complicated application scenarios.Comment: 14 page
    • …
    corecore