22 research outputs found

    On-line Handwritten Character Recognition: An Implementation of Counterpropagation Neural Net

    Get PDF
    On-line handwritten scripts are usually dealt with pen tip traces from pen-down to pen-up positions. Time evaluation of the pen coordinates is also considered along with trajectory information. However, the data obtained needs a lot of preprocessing including filtering, smoothing, slant removing and size normalization before recognition process. Instead of doing such lengthy preprocessing, this paper presents a simple approach to extract the useful character information. This work evaluates the use of the counter- propagation neural network (CPN) and presents feature extraction mechanism in full detail to work with on-line handwriting recognition. The obtained recognition rates were 60% to 94% using the CPN for different sets of character samples. This paper also describes a performance study in which a recognition mechanism with multiple hresholds is evaluated for counter-propagation architecture. The results indicate that the application of multiple thresholds has significant effect on recognition mechanism. The method is applicable for off-line character recognition as well. The technique is tested for upper-case English alphabets for a number of different styles from different peoples

    Efficient Yet Deep Convolutional Neural Networks for Semantic Segmentation

    Full text link
    Semantic Segmentation using deep convolutional neural network pose more complex challenge for any GPU intensive task. As it has to compute million of parameters, it results to huge memory consumption. Moreover, extracting finer features and conducting supervised training tends to increase the complexity. With the introduction of Fully Convolutional Neural Network, which uses finer strides and utilizes deconvolutional layers for upsampling, it has been a go to for any image segmentation task. In this paper, we propose two segmentation architecture which not only needs one-third the parameters to compute but also gives better accuracy than the similar architectures. The model weights were transferred from the popular neural net like VGG19 and VGG16 which were trained on Imagenet classification data-set. Then we transform all the fully connected layers to convolutional layers and use dilated convolution for decreasing the parameters. Lastly, we add finer strides and attach four skip architectures which are element-wise summed with the deconvolutional layers in steps. We train and test on different sparse and fine data-sets like Pascal VOC2012, Pascal-Context and NYUDv2 and show how better our model performs in this tasks. On the other hand our model has a faster inference time and consumes less memory for training and testing on NVIDIA Pascal GPUs, making it more efficient and less memory consuming architecture for pixel-wise segmentation.Comment: 8 page

    Experiments on the Application of IOHMMs to Model Financial Returns Series

    Get PDF
    Input/Output Hidden Markov Models (IOHMMs) are conditional hidden Markov models in which the emission (and possibly the transition) probabilities can be conditioned on an input sequence. For example, these conditional distributions can be linear, logistic, or non-linear (using for example multi-layer neural networks). We compare the generalization performance of several models which are special cases of Input/Output Hidden Markov Models on financial time-series prediction tasks: an unconditional Gaussian, a conditional linear Gaussian, a mixture of Gaussians, a mixture of conditional linear Gaussians, a hidden Markov model, and various IOHMMs. The experiments compare these models on predicting the conditional density of returns of market and sector indices. Note that the unconditional Gaussian estimates the first moment with the historical average. The results show that, although for the first moment the historical average gives the best results, for the higher moments, the IOHMMs yielded significantly better performance, as estimated by the out-of-sample likelihood. "Input/Output Hidden Markov Models" (IOHMMs) sont des modèles de Markov cachés pour lesquels les probabilités d'émission (et possiblement de transition) peuvent dépendre d'une séquence d'entrée. Par exemple, ces distributions conditionnelles peuvent être linéaires, logistique, ou non-linéaire (utilisant, par exemple, une réseau de neurones multi-couches). Nous comparons les performances de généralisation de plusieurs modèles qui sont des cas particuliers de IOHMMs pour des problèmes de prédictions de séries financières : une gaussienne inconditionnelle, une gaussienne linéaire conditionnelle, une mixture de gaussienne, une mixture de gaussiennes linéaires conditionnelles, un modèle de Markov caché, et divers IOHMMs. Les expériences comparent ces modèles sur leur prédictions de la densité conditionnelle des rendements des indices sectoriels et du marché. Notons qu'une gaussienne inconditionnelle estime le premier moment avec une moyenne historique. Les résultats montrent que, même si la moyenne historique donne les meilleurs résultats pour le premier moment, pour les moments d'ordres supérieurs les IOHMMs performent significativement mieux, comme estimé par la vraisemblance hors-échantillon.Input/Output Hidden Markov Model (IOHMM), financial series, volatility, Modèles de Markov cachés, IOHMM, séries financières, volatilité

    Fully Convolutional Networks for Semantic Segmentation

    Full text link
    Convolutional networks are powerful visual models that yield hierarchies of features. We show that convolutional networks by themselves, trained end-to-end, pixels-to-pixels, exceed the state-of-the-art in semantic segmentation. Our key insight is to build "fully convolutional" networks that take input of arbitrary size and produce correspondingly-sized output with efficient inference and learning. We define and detail the space of fully convolutional networks, explain their application to spatially dense prediction tasks, and draw connections to prior models. We adapt contemporary classification networks (AlexNet, the VGG net, and GoogLeNet) into fully convolutional networks and transfer their learned representations by fine-tuning to the segmentation task. We then define a novel architecture that combines semantic information from a deep, coarse layer with appearance information from a shallow, fine layer to produce accurate and detailed segmentations. Our fully convolutional network achieves state-of-the-art segmentation of PASCAL VOC (20% relative improvement to 62.2% mean IU on 2012), NYUDv2, and SIFT Flow, while inference takes one third of a second for a typical image.Comment: to appear in CVPR (2015

    On the Use of Attention Mechanism in a Seq2Seq based Approach for Off-line Handwritten Digit String Recognition

    Get PDF
    International audienceIn this work, we investigate the use of the attention mechanism in deep learning for a better reading of handwritten digit strings in digitized images. The proposed recognition system built upon a CNN (Convolutional Neural Network) and two RNNs (Recurrent Neural Networks), acting as Encoder and Decoder and using the attention mechanism. We used a 1D mechanism for attention location with a "soft" alignment attention which has the peculiarity of having an easily calculable gradient and thus to integrate well with the network. Experimental results on data from ORAND-CAR A, ORAND-CAR B and CVL HDS databases compare favorably to other published methods
    corecore