557 research outputs found

    FEATURE EXTRACTION AND RECOGNITION ON TRAFFIC SIGN IMAGES

    Get PDF
    FEATURE EXTRACTION AND RECOGNITION ON TRAFFIC SIGN IMAGESAbstractIt is vital that the traffic signs used to ensure the order of the traffic are perceived by the drivers. Traffic signs have international standards that allow the driver to learn about the road and the environment while driving. Traffic sign recognition systems have recently started to be used in vehicles in order to improve traffic safety. Machine learning methods are used in the field of image recognition. Deep learning methods increase the classification success by extracting the hidden and interesting features in the image. Images contain many features and this situation can affect success in classification problems. It can also reveal the need for high-capacity hardware. In order to solve these problems, convolutional neural networks can be used to extract meaningful features from the image. In this study, we created a dataset containing 1500 images of 14 different traffic signs that are frequently used on Turkey highways. The features of the images in this dataset were extracted using convolutional neural networks from deep learning architectures. The 1000 features obtained were classified using the Random Forest method from machine learning algorithms. 93.7% success was achieved as a result of this classification process.Keywords: Classification, Convolution neural network, Feature extraction, Random forest, Traffic signsTRAFİK İŞARETİ GÖRÜNTÜLERİNDE ÖZELLİK ÇIKARMA VE TANIMAÖzetTrafiğin düzenini sağlamak amacıyla kullanılan trafik levhalarını sürücülerin algılaması hayati önem taşımaktadır. Sürüş esnasında sürücünün yol ve çevre hakkında bilgi edinebilmesini sağlayan trafik levhaları uluslararası standartlara sahiptir. Trafik levhası tanıma sistemleri son zamanlarda trafik güvenliğini arttırmak amacıyla araçlarda kullanılmaya başlamıştır. Makine öğrenmesi yöntemleri görüntü tanıma alanında kullanılmaktadır.  Derin öğrenme yöntemleri, görüntüde yer alan gizli ve ilginç özellikleri çıkarak sınıflandırma başarısını arttırmaktadır. Görüntüler çok sayıda özellik içermektedir ve bu durum sınıflandırma problemlerinde başarıyı etkileyebilmektedir. Ayrıca yüksek kapasiteli donanım gereksinimini de ortaya çıkarabilmektedir. Bu sorunların çözülebilmesi için görüntüden anlamlı özelliklerin çıkarılmasında konvolüsyonel sinir ağları kullanılabilmektedir. Bu çalışmada Türkiye’deki karayollarında sıklıkla kullanılan 14 farklı trafik levhasına ait 1500 görüntü içeren bir veriseti tarafımızca oluşturulmuştur. Bu veriseti kullanılarak derin öğrenme mimarilerinden konvolüsyonel sinir ağları kullanılarak görüntülerin özellikleri çıkarılmıştır. Elde edilen 1000 özellik makine öğrenmesi algoritmalarından Random Forest yöntemi kullanılarak sınıflandırılmıştır. Bu sınıflandırma işlemi sonucunda %93.7 başarı elde edilmiştir.Anahtar Kelimeler: Konvolüsyonel sinir ağları, Özellik çıkarma, Random forest, Sınıflandırma, Trafik işaretleri

    Multi-scale Discriminant Saliency with Wavelet-based Hidden Markov Tree Modelling

    Full text link
    The bottom-up saliency, an early stage of humans' visual attention, can be considered as a binary classification problem between centre and surround classes. Discriminant power of features for the classification is measured as mutual information between distributions of image features and corresponding classes . As the estimated discrepancy very much depends on considered scale level, multi-scale structure and discriminant power are integrated by employing discrete wavelet features and Hidden Markov Tree (HMT). With wavelet coefficients and Hidden Markov Tree parameters, quad-tree like label structures are constructed and utilized in maximum a posterior probability (MAP) of hidden class variables at corresponding dyadic sub-squares. Then, a saliency value for each square block at each scale level is computed with discriminant power principle. Finally, across multiple scales is integrated the final saliency map by an information maximization rule. Both standard quantitative tools such as NSS, LCC, AUC and qualitative assessments are used for evaluating the proposed multi-scale discriminant saliency (MDIS) method against the well-know information based approach AIM on its released image collection with eye-tracking data. Simulation results are presented and analysed to verify the validity of MDIS as well as point out its limitation for further research direction.Comment: arXiv admin note: substantial text overlap with arXiv:1301.396

    Deep Logo Authenticity: Leveraging R-CNN for Counterfeit Logo Detection in E-commerce

    Get PDF
    In the rapidly evolving realm of electronic commerce, ensuring the accuracy and authenticity of merchandise assumes paramount importance in maintaining consumer trust and platform reliability. One of the prominent challenges encountered within this particular domain revolves around the pervasive prevalence of counterfeit products, often discernible through subtle deviations in brand insignias. This research paper introduces a novel approach to detect counterfeit logos on electronic commerce platforms using Region-based Convolutional Neural Networks (R-CNN). Traditional approaches often rely on manual verification or basic image comparisons, both of which have drawbacks in terms of scalability and consistent accuracy. The methodology utilized in our research capitalizes on the capabilities of deep learning algorithms to precisely identify and classify logos depicted in product images, proficiently distinguishing genuine logos from counterfeit ones with a significant degree of precision. A meticulously curated dataset was compiled, encompassing genuine and counterfeit logos sourced from renowned brands. By means of intensive training, our model demonstrated remarkable aptitude, surpassing the capabilities of contemporary methodologies. The current investigation not only offers a significant contribution to enhancing the security and reliability of electronic commerce platforms, but also establishes the foundation for the advancement of advanced counterfeit detection methodologies within the domain of digital marketplaces

    Hidden Markov Models

    Get PDF
    Hidden Markov Models (HMMs), although known for decades, have made a big career nowadays and are still in state of development. This book presents theoretical issues and a variety of HMMs applications in speech recognition and synthesis, medicine, neurosciences, computational biology, bioinformatics, seismology, environment protection and engineering. I hope that the reader will find this book useful and helpful for their own research

    Investigating the latency cost of statistical learning of a Gaussian mixture simulating on a convolutional density network with adaptive batch size technique for background modeling

    Get PDF
    Background modeling is a promising field of study in video analysis, with a wide range of applications in video surveillance. Deep neural networks have proliferated in recent years as a result of effective learning-based approaches to motion analysis. However, these strategies only provide a partial description of the observed scenes' insufficient properties since they use a single-valued mapping to estimate the target background's temporal conditional averages. On the other hand, statistical learning in the imagery domain has become one of the most widely used approaches due to its high adaptability to dynamic context transformation, especially Gaussian Mixture Models. Specifically, these probabilistic models aim to adjust latent parameters to gain high expectation of realistically observed data; however, this approach only concentrates on contextual dynamics in short-term analysis. In a prolonged investigation, it is challenging so that statistical methods cannot reserve the generalization of long-term variation of image data. Balancing the trade-off between traditional machine learning models and deep neural networks requires an integrated approach to ensure accuracy in conception while maintaining a high speed of execution. In this research, we present a novel two-stage approach for detecting changes using two convolutional neural networks in this work. The first architecture is based on unsupervised Gaussian mixtures statistical learning, which is used to classify the salient features of scenes. The second one implements a light-weighted pipeline of foreground detection. Our two-stage system has a total of approximately 3.5K parameters but still converges quickly to complex motion patterns. Our experiments on publicly accessible datasets demonstrate that our proposed networks are not only capable of generalizing regions of moving objects with promising results in unseen scenarios, but also competitive in terms of performance quality and effectiveness foreground segmentation. Apart from modeling the data's underlying generator as a non-convex optimization problem, we briefly examine the communication cost associated with the network training by using a distributed scheme of data-parallelism to simulate a stochastic gradient descent algorithm with communication avoidance for parallel machine learnin

    A survey on deep learning techniques for image and video semantic segmentation

    Get PDF
    Image semantic segmentation is more and more being of interest for computer vision and machine learning researchers. Many applications on the rise need accurate and efficient segmentation mechanisms: autonomous driving, indoor navigation, and even virtual or augmented reality systems to name a few. This demand coincides with the rise of deep learning approaches in almost every field or application target related to computer vision, including semantic segmentation or scene understanding. This paper provides a review on deep learning methods for semantic segmentation applied to various application areas. Firstly, we formulate the semantic segmentation problem and define the terminology of this field as well as interesting background concepts. Next, the main datasets and challenges are exposed to help researchers decide which are the ones that best suit their needs and goals. Then, existing methods are reviewed, highlighting their contributions and their significance in the field. We also devote a part of the paper to review common loss functions and error metrics for this problem. Finally, quantitative results are given for the described methods and the datasets in which they were evaluated, following up with a discussion of the results. At last, we point out a set of promising future works and draw our own conclusions about the state of the art of semantic segmentation using deep learning techniques.This work has been funded by the Spanish Government TIN2016-76515-R funding for the COMBAHO project, supported with Feder funds. It has also been supported by a Spanish national grant for PhD studies FPU15/04516 (Alberto Garcia-Garcia). In addition, it was also funded by the grant Ayudas para Estudios de Master e Iniciacion a la Investigacion from the University of Alicante

    A Multicamera System for Gesture Tracking With Three Dimensional Hand Pose Estimation

    Get PDF
    The goal of any visual tracking system is to successfully detect then follow an object of interest through a sequence of images. The difficulty of tracking an object depends on the dynamics, the motion and the characteristics of the object as well as on the environ ment. For example, tracking an articulated, self-occluding object such as a signing hand has proven to be a very difficult problem. The focus of this work is on tracking and pose estimation with applications to hand gesture interpretation. An approach that attempts to integrate the simplicity of a region tracker with single hand 3D pose estimation methods is presented. Additionally, this work delves into the pose estimation problem. This is ac complished by both analyzing hand templates composed of their morphological skeleton, and addressing the skeleton\u27s inherent instability. Ligature points along the skeleton are flagged in order to determine their effect on skeletal instabilities. Tested on real data, the analysis finds the flagging of ligature points to proportionally increase the match strength of high similarity image-template pairs by about 6%. The effectiveness of this approach is further demonstrated in a real-time multicamera hand tracking system that tracks hand gestures through three-dimensional space as well as estimate the three-dimensional pose of the hand
    corecore