196,709 research outputs found

    Learning from Long-Tailed Noisy Data with Sample Selection and Balanced Loss

    Full text link
    The success of deep learning depends on large-scale and well-curated training data, while data in real-world applications are commonly long-tailed and noisy. Many methods have been proposed to deal with long-tailed data or noisy data, while a few methods are developed to tackle long-tailed noisy data. To solve this, we propose a robust method for learning from long-tailed noisy data with sample selection and balanced loss. Specifically, we separate the noisy training data into clean labeled set and unlabeled set with sample selection, and train the deep neural network in a semi-supervised manner with a novel balanced loss based on model bias. Experiments on benchmarks demonstrate that our method outperforms existing state-of-the-art methods

    Application of neural network to study share price volatility.

    Get PDF
    by Lam King Wan.Thesis (M.B.A.)--Chinese University of Hong Kong, 1999.Includes bibliographical references (leaves 72-73).ABSTRACT --- p.ii.TABLE OF CONTENTS --- p.iv.SectionChapter I. --- OBJECTIVE --- p.1Chapter II. --- INTRODUCTION --- p.3The principal investment risk --- p.3Effect of risk on investment --- p.4Investors' concern for investment risk --- p.6Chapter III. --- THE INPUT PARAMETERS --- p.9Chapter IV. --- LITERATURE REVIEW --- p.15What is an artificial neural network? --- p.15What is a neuron? --- p.16Biological versus artificial neuron --- p.16Operation of a neural network --- p.17Neural network paradigm --- p.20Feedforward as the most suitable form of neural network --- p.22Capability of neural network --- p.23The learning process --- p.25Testing the network --- p.29Neural network computing --- p.29Neural network versus conventional computer --- p.30Neural network versus a knowledge based system --- p.32Strength of neural network --- p.34Weaknesses of neural network --- p.35Chapter V. --- NEURAL NETWORK AS A TOOL FOR INVESTMENT ANALYSIS --- p.38Neural network in financial applications --- p.38Trading in the stock market --- p.41Why neural network could outperform in the stock market? --- p.43Applications of neural network --- p.45Chapter VI. --- BUILDING THE NEURAL NETWORK MODEL --- p.47Implementation process --- p.48Step 1´ؤ Problem specification --- p.49Step 2 ´ؤ Data collection --- p.51Step 3 ´ؤ Data analysis and transformation --- p.55Step 4 ´ؤ Training data set extraction --- p.58Step 5 ´ؤ Selection of network architecture --- p.60Step 6 ´ؤ Selection of training algorithm --- p.62Step 7 ´ؤ Training the network --- p.64Step 8 ´ؤ Model deployment --- p.65Chapter 7 --- RESULT AND CONCLUSION --- p.67Result --- p.67Conclusion --- p.69BIBLIOGRAPHY --- p.7

    Multi-criteria Evolution of Neural Network Topologies: Balancing Experience and Performance in Autonomous Systems

    Full text link
    Majority of Artificial Neural Network (ANN) implementations in autonomous systems use a fixed/user-prescribed network topology, leading to sub-optimal performance and low portability. The existing neuro-evolution of augmenting topology or NEAT paradigm offers a powerful alternative by allowing the network topology and the connection weights to be simultaneously optimized through an evolutionary process. However, most NEAT implementations allow the consideration of only a single objective. There also persists the question of how to tractably introduce topological diversification that mitigates overfitting to training scenarios. To address these gaps, this paper develops a multi-objective neuro-evolution algorithm. While adopting the basic elements of NEAT, important modifications are made to the selection, speciation, and mutation processes. With the backdrop of small-robot path-planning applications, an experience-gain criterion is derived to encapsulate the amount of diverse local environment encountered by the system. This criterion facilitates the evolution of genes that support exploration, thereby seeking to generalize from a smaller set of mission scenarios than possible with performance maximization alone. The effectiveness of the single-objective (optimizing performance) and the multi-objective (optimizing performance and experience-gain) neuro-evolution approaches are evaluated on two different small-robot cases, with ANNs obtained by the multi-objective optimization observed to provide superior performance in unseen scenarios

    Learning icons appearance similarity

    Get PDF
    Selecting an optimal set of icons is a crucial step in the pipeline of visual design to structure and navigate through content. However, designing the icons sets is usually a difficult task for which expert knowledge is required. In this work, to ease the process of icon set selection to the users, we propose a similarity metric which captures the properties of style and visual identity. We train a Siamese Neural Network with an online dataset of icons organized in visually coherent collections that are used to adaptively sample training data and optimize the training process. As the dataset contains noise, we further collect human-rated information on the perception of icon's similarity which will be used for evaluating and testing the proposed model. We present several results and applications based on searches, kernel visualizations and optimized set proposals that can be helpful for designers and non-expert users while exploring large collections of icons.Comment: 12 pages, 11 figure

    Neural Network Based Inferential Model For Ethane Steam Cracking Furnace

    Get PDF
    The product yield distribution of ethane steam cracking is typically obtained using analysers and lab sampling. Since both methods take time to produce results, primarily depending on them to determine main product yield will hinder immediate control action on the process. In order to resolve this issue, an inferential sensor is required. In this study, a neural network based inferential model is developed. The ethane steam cracking process has been modelled using ASPEN Plus and validated with industrial data taken from literature. The relative error (RE) of the model outputs obtained are less than 10%. The ASPEN Plus model is used for input variable selection, nonlinearity assessment, and data generation for neural network modelling. The input variable selection study found that five variables are significantly influential to the ethane and ethylene yields, namely reactor pressure, coil outlet temperature, steam-hydrocarbon ratio, feed composition, and fuel composition. Nonlinearity assessment of the process shows that the process exhibit asymmetrical response and input multiplicities characteristics, and thus, can be classified as a nonlinear process. Data generated from the ASPEN Plus model is used for training, validation, and testing. Two methods have been used to generate the data which are sequential excitation and simultaneous excitation. Four variables are individually excited and combined to make a sequential excitation profile. Data from sequential excitation is divided into training and validation while data from simultaneous excitation is used solely for testing. Three neural network model, namely the Feedforward Neural Network (FFNN), the Generalized Regression Neural Network (GRNN), and the Extreme Learning Machine Neural Network (ELM-NN) are developed and they are evaluated in terms of prediction accuracy and computational time. The evaluation results show that ELM-NN prediction accuracy is higher than FFNN and GRNN. To train, the best model for ELM-NN, GRNN, and FFNN models require 0.0068 seconds, 0.35 seconds, and 12 seconds respectively. In terms of computation time of new set of input data sample, all three models require less than 0.05 seconds to compute one sample of data. However, computation time of the trained GRNN model increases exponentially with the increasing amount of data samples in a batch while for trained FFNN and trained ELM-NN model, the increment is not significant. Out of the three models, the ELM-NN gives the best performance in terms of prediction accuracy and computational time. The R2 of the ELM-NN model is 91.3% and 82.6% for ethane and ethylene yield respectively. The model requires 0.0068 seconds to train and 0.0001 seconds to compute ethane yield and ethylene yields from a new set of input data. This makes the model suitable for applications in real time inferential control system

    Would credit scoring work for Islamic finance? A neural network approach

    Get PDF
    Purpose – The main aim of this paper is to distinguish whether the decision making process of the Islamic financial houses in the UK can be improved through the use of credit scoring modeling techniques as opposed to the currently used judgmental approaches. Subsidiary aims are to identify how scoring models can reclassify accepted applicants who later are considered as having bad credit and how many of the rejected applicants are later considered as having good credit; and highlight significant variables that are crucial in terms of accepting and rejecting applicants which can further aid the decision making process. Design/methodology/approach – A real data-set of 487 applicants are used consisting of 336 accepted credit applications and 151 rejected credit applications make to an Islamic finance house in the UK. In order to build the proposed scoring models, the data-set is divided into training and hold-out sub-set. The training sub-set is used to build the scoring models and the hold-out sub-set is used to test the predictive capabilities of the scoring models.70 percent of the overall applicants will be used for the training sub-set and 30 percent will be used for the testing sub-set. Three statistical modeling techniques namely Discriminant Analysis (DA), Logistic Regression (LR) and Multi-layer Perceptron (MP) neural network are used to build the proposed scoring models. Findings – Our findings reveal that the LR model has the highest Correct Classification (CC) rate in the training sub-set whereas MP outperforms other techniques and has the highest CC rate in the hold-out sub-set. MP also outperforms other techniques in terms of predicting the rejected credit applications and has the lowest Misclassification Cost (MC) above other techniques. In addition, results from MP models show that monthly expenses, age and marital status are identified as the key factors affecting the decision making process. Research limitations/implications – Although our sample is small and restricted to an Islamic Finance house in the UK the results are robust. Future research could consider enlarging the sample in the UK and also internationally allowing for cultural differences to be identified. The results indicate that the scoring models can be of great benefit to Islamic finance houses in regards to their decision making processes of accepting and rejecting new credit applications and thus improve their efficiency and effectiveness. Originality/value –Our contribution is the first to apply credit scoring modeling techniques in Islamic Finance. Also in building a scoring model our application applies a different approach by using accepted and rejected credit applications instead of good and bad credit histories. This identifies opportunity costs of misclassifying credit applications as rejected

    A Very Brief Introduction to Machine Learning With Applications to Communication Systems

    Get PDF
    Given the unprecedented availability of data and computing resources, there is widespread renewed interest in applying data-driven machine learning methods to problems for which the development of conventional engineering solutions is challenged by modelling or algorithmic deficiencies. This tutorial-style paper starts by addressing the questions of why and when such techniques can be useful. It then provides a high-level introduction to the basics of supervised and unsupervised learning. For both supervised and unsupervised learning, exemplifying applications to communication networks are discussed by distinguishing tasks carried out at the edge and at the cloud segments of the network at different layers of the protocol stack

    A selective learning method to improve the generalization of multilayer feedforward neural networks.

    Get PDF
    Multilayer feedforward neural networks with backpropagation algorithm have been used successfully in many applications. However, the level of generalization is heavily dependent on the quality of the training data. That is, some of the training patterns can be redundant or irrelevant. It has been shown that with careful dynamic selection of training patterns, better generalization performance may be obtained. Nevertheless, generalization is carried out independently of the novel patterns to be approximated. In this paper, we present a learning method that automatically selects the training patterns more appropriate to the new sample to be predicted. This training method follows a lazy learning strategy, in the sense that it builds approximations centered around the novel sample. The proposed method has been applied to three different domains: two artificial approximation problems and a real time series prediction problem. Results have been compared to standard backpropagation using the complete training data set and the new method shows better generalization abilities.Publicad
    corecore