3,466 research outputs found

    Applying MDL to Learning Best Model Granularity

    Get PDF
    The Minimum Description Length (MDL) principle is solidly based on a provably ideal method of inference using Kolmogorov complexity. We test how the theory behaves in practice on a general problem in model selection: that of learning the best model granularity. The performance of a model depends critically on the granularity, for example the choice of precision of the parameters. Too high precision generally involves modeling of accidental noise and too low precision may lead to confusion of models that should be distinguished. This precision is often determined ad hoc. In MDL the best model is the one that most compresses a two-part code of the data set: this embodies ``Occam's Razor.'' In two quite different experimental settings the theoretical value determined using MDL coincides with the best value found experimentally. In the first experiment the task is to recognize isolated handwritten characters in one subject's handwriting, irrespective of size and orientation. Based on a new modification of elastic matching, using multiple prototypes per character, the optimal prediction rate is predicted for the learned parameter (length of sampling interval) considered most likely by MDL, which is shown to coincide with the best value found experimentally. In the second experiment the task is to model a robot arm with two degrees of freedom using a three layer feed-forward neural network where we need to determine the number of nodes in the hidden layer giving best modeling performance. The optimal model (the one that extrapolizes best on unseen examples) is predicted for the number of nodes in the hidden layer considered most likely by MDL, which again is found to coincide with the best value found experimentally.Comment: LaTeX, 32 pages, 5 figures. Artificial Intelligence journal, To appea

    Performance analysis of Handwritten Devnagari Character Recognition using Feed Forward , Radial Basis , Elman Back Propagation, and Pattern Recognition Neural Network Model Using Different Feature Extraction Methods

    Get PDF
    This paper describes the performance analysis for the four types of neural network with different feature extraction methods for character recognition of hand written devnagari alphabets. We have implemented four types of networks i.e. Feed forward , Radial basis, Elman back propagation and Pattern recognition neural network using three different types of feature extraction methods i.e. pixel value, histogram and blocks mean for each network. These algorithms have been performed better than the conventional approaches of neural network for pattern recognition. It has been analyzed that the Radial Basis neural network performs better compared to other types of networks

    A new hybrid convolutional neural network and eXtreme gradient boosting classifier for recognizing handwritten Ethiopian characters

    Get PDF
    Handwritten character recognition has been profoundly studied for many years in the field of pattern recognition. Due to its vast practical applications and financial implications, handwritten character recognition is still an important research area. In this research, the Handwritten Ethiopian Character Recognition (HECR) dataset has been prepared to train the model. The images in the HECR dataset were organized with more than one color pen RGB main spaces that have been size normalized to 28 Ă— 28 pixels. The dataset is a combination of scripts (Fidel in Ethiopia), numerical representations, punctuations, tonal symbols, combining symbols, and special characters. These scripts have been used to write ancient histories, science, and arts of Ethiopia and Eritrea. In this study, a hybrid model of two super classifiers: Convolutional Neural Network (CNN) and eXtreme Gradient Boosting (XGBoost) is proposed for classification. In this integrated model, CNN works as a trainable automatic feature extractor from the raw images and XGBoost takes the extracted features as an input for recognition and classification. The output error rates of the hybrid model and CNN with a fully connected layer are compared. A 0.4630 and 0.1612 error rates are achieved in classifying the handwritten testing dataset images, respectively. Thus XGBoost as a classifier performs a better result than the traditional fully connected layer

    Adaptive restoration of text images containing touching and broken characters

    Full text link
    For document processing systems, automated data entry is generally performed by optical character recognition (OCR) systems. To make these systems practical, reliable OCR systems are essential. However, distortions in document images cause character recognition errors, thereby, reducing the accuracy of OCR systems. In document images, most OCR errors are caused by broken and touching characters. This thesis presents an adaptive system to restore text images distorted by touching and broken characters. The adaptive system uses the distorted text image and the output from an OCR system to generate the training character image. Using the training image and the distorted image, the system trains an adaptive restoration filter and then uses the trained filter to restore the distorted text image. To demonstrate the performance of this technique, it was applied to several distorted images containing touching or broken characters. The results show that this technique can improve both pixel and OCR accuracy of distorted text images containing touching or broken characters

    Performance Analysis of Handwritten Marathi Character Recognition with RBF, Cascade, Elman and Feed Forward Neural Networks

    Get PDF
    Character recognition of handwritten Marathi curve scripts is one of the most challenging areas of research in neural networks due to high variability in writing styles. Marathi characters have shirolekhas and spines. This seriously affects many of the performance recognition parameters and much more.In this paper, we are performing the performance analysis of RBF neural network, Cascade Neural network, Elman Neural network and Feed forward neural network for the character recognition of handwritten Marathi curve scripts. For the experiment, we have taken in to account the six samples each of 48 Marathi characters. For every sampled character, the �Edge detection and dilation method of Feature extraction�with a set of image pre-processing operations have been performed. Here to study and analyze the performance of these four neural networks, firstly we have created the network, trained the network, simulated the network and plotted the regression plots. It has been analyzed that RBF neural networks has a high regression value as compared to the rest of the methods for the training set

    Design compact and efficient recurrent neural networks for natural language processing tasks

    Get PDF
    The present work takes into account the compactness and efficiency of Recurrent Neural Networks (RNNs) for solving Natural Language Processing (NLP) tasks. RNNs are a class of Artificial Neural Networks (ANNs). Compared to Feed-forward Neural Networks (FNNs), RNN architecture is cyclic, i.e. the connection between nodes form cycles. This subtle difference has actually a huge impact on solving sequence-based problems, e.g. NLP tasks.In particular, the first advantage of RNNs regards their ability to modellong-range time dependencies, which is a very desirable property for natural languagedata, where word’s meaning is highly dependent on its context. The second advantage of RNNs is that are flexible and accept as input many different datatypes and representation. This is again the case of natural language data, whichcan come in different sizes, e.g. words with different lengths, and types, e.g. sequences or trees.ope
    • …
    corecore