475,473 research outputs found

    A robust algorithm for detection and classification of traffic signs in video data

    Get PDF
    —The accurate identification and recognition of the traffic signs is a challenging problem as the developed systems have to address a large number of imaging problems such as motion artifacts, various weather conditions, shadows and partial occlusion, issues that are often encountered in video traffic sequences that are captured from a moving vehicle. These factors substantially degrade the performance of the existing traffic sign recognition (TSR) systems and in this paper we detail the implementation of a new strategy that entails three distinct computational stages. The first component addresses the robust identification of the candidate traffic signs in each frame of the video sequence. The second component discards the traffic sign candidates that do not comply with stringent shape constraints, and the last component implements the classification of the traffic signs using Support Vector Machines (SVMs). The main novel elements of our TSR algorithm are given by the approach that has been developed for traffic sign classification and by the experimental evaluation that was employed to identify the optimal image attributes that are able to maximize the traffic sign classification performance. The TSR algorithm has been validated using video sequences that include the most important categories of signs that are used to regulate the traffic on the Irish and UK roads, and it achieved 87.6% sign detection, 99.2% traffic sign classification accuracy and 86.7% overall traffic sign recognition

    Imaging-in-flow: digital holographic microscopy as a novel tool to detect and classify nanoplanktonic organisms

    Get PDF
    Traditional taxonomic identification of planktonic organisms is based on light microscopy, which is both time-consuming and tedious. In response, novel ways of automated (machine) identification, such as flow cytometry, have been investigated over the last two decades. To improve the taxonomic resolution of particle analysis, recent developments have focused on "imaging-in-flow," i.e., the ability to acquire microscopic images of planktonic cells in a flow-through mode. Imaging-in-flow systems are traditionally based on classical brightfield microscopy and are faced with a number of issues that decrease the classification performance and accuracy (e. g., projection variance of cells, migration of cells out of the focus plane). Here, we demonstrate that a combination of digital holographic microscopy (DHM) with imaging-in-flow can improve the detection and classification of planktonic organisms. In addition to light intensity information, DHM provides quantitative phase information, which generates an additional and independent set of features that can be used in classification algorithms. Moreover, the capability of digitally refocusing greatly increases the depth of field, enables a more accurate focusing of cells, and reduces the effects of position variance. Nanoplanktonic organisms similar in shape were successfully classified from images captured with an off-axis DHM with partial coherence. Textural features based on DHM phase information proved more efficient in separating the three tested phytoplankton species compared with shape-based features or textural features based on light intensity. An overall classification score of 92.4% demonstrates the potential of holographic-based imaging-in-flow for similar looking organisms in the nanoplankton range

    Radar and RGB-depth sensors for fall detection: a review

    Get PDF
    This paper reviews recent works in the literature on the use of systems based on radar and RGB-Depth (RGB-D) sensors for fall detection, and discusses outstanding research challenges and trends related to this research field. Systems to detect reliably fall events and promptly alert carers and first responders have gained significant interest in the past few years in order to address the societal issue of an increasing number of elderly people living alone, with the associated risk of them falling and the consequences in terms of health treatments, reduced well-being, and costs. The interest in radar and RGB-D sensors is related to their capability to enable contactless and non-intrusive monitoring, which is an advantage for practical deployment and users’ acceptance and compliance, compared with other sensor technologies, such as video-cameras, or wearables. Furthermore, the possibility of combining and fusing information from The heterogeneous types of sensors is expected to improve the overall performance of practical fall detection systems. Researchers from different fields can benefit from multidisciplinary knowledge and awareness of the latest developments in radar and RGB-D sensors that this paper is discussing

    Exploiting Deep Features for Remote Sensing Image Retrieval: A Systematic Investigation

    Full text link
    Remote sensing (RS) image retrieval is of great significant for geological information mining. Over the past two decades, a large amount of research on this task has been carried out, which mainly focuses on the following three core issues: feature extraction, similarity metric and relevance feedback. Due to the complexity and multiformity of ground objects in high-resolution remote sensing (HRRS) images, there is still room for improvement in the current retrieval approaches. In this paper, we analyze the three core issues of RS image retrieval and provide a comprehensive review on existing methods. Furthermore, for the goal to advance the state-of-the-art in HRRS image retrieval, we focus on the feature extraction issue and delve how to use powerful deep representations to address this task. We conduct systematic investigation on evaluating correlative factors that may affect the performance of deep features. By optimizing each factor, we acquire remarkable retrieval results on publicly available HRRS datasets. Finally, we explain the experimental phenomenon in detail and draw conclusions according to our analysis. Our work can serve as a guiding role for the research of content-based RS image retrieval

    What is Holding Back Convnets for Detection?

    Full text link
    Convolutional neural networks have recently shown excellent results in general object detection and many other tasks. Albeit very effective, they involve many user-defined design choices. In this paper we want to better understand these choices by inspecting two key aspects "what did the network learn?", and "what can the network learn?". We exploit new annotations (Pascal3D+), to enable a new empirical analysis of the R-CNN detector. Despite common belief, our results indicate that existing state-of-the-art convnet architectures are not invariant to various appearance factors. In fact, all considered networks have similar weak points which cannot be mitigated by simply increasing the training data (architectural changes are needed). We show that overall performance can improve when using image renderings for data augmentation. We report the best known results on the Pascal3D+ detection and view-point estimation tasks
    • 

    corecore