526 research outputs found

    Methods for Ellipse Detection from Edge Maps of Real Images

    Get PDF

    Age Differences in Vestibular Processing: Neural and Behavioral Evidence

    Full text link
    The vestibular system is well known for its role in balance, but its mechanisms of action in this role are not well understood. My dissertation aims to provide a better understanding of vestibular brain function, its correlation with postural control, and its alteration with advancing age. This is an important topic considering that falls are the current leading cause of injuries in older adults in the U.S., and they have negative consequences on wellbeing and independence. In this dissertation, I first review the conventional methods for studying vestibular function in the human brain, and I evaluate a novel MRI-compatible method, which relies on a pneumatic tapper. This approach successfully induces vestibular responses, while preventing the aversive effects of stimulation that are common in other approaches. Next, I assess age differences in brain responses to pneumatic vestibular stimulation, and find that older adults demonstrate less sensitivity to stimulation. Also, those with better postural control exhibit less deactivation of cross-modal sensory regions (e.g. visual and somatosensory cortices). This greater engagement of non-vestibular sensory regions in older adults with better balance could be a mechanism to compensate for inefficient vestibular processing. Consistent with this hypothesis, the relationship between postural control and deactivation of sensory regions was only evident in tasks of low difficulty (i.e. normal stance) in which compensatory neural recruitment might be most effective. After assessing the brain responses to vestibular stimulation in terms of activation and deactivation, I examine connectivity of the vestibular cortex with other regions. This last experiment demonstrates that vestibular cortex connectivity increases in response to vestibular stimulation, and young adults exhibit greater connectivity relative to older adults. Also, connectivity predicts postural stability in high difficulty tasks for young adults, and in low difficulty tasks for older adults. Better balance in young adults is associated with less vestibular connectivity (i.e. they engaged vestibular cortex more selectively), whereas better balance in older adults is associated with higher connectivity (i.e. more recruitment of other sensory regions). These findings reinforce the conclusions from the second experiment, and provide more evidence in support of the compensation related utilization of neural circuits hypothesis (CRUNCH) of neural processing in older adults.PHDKines & Psychology PhDUniversity of Michigan, Horace H. Rackham School of Graduate Studieshttps://deepblue.lib.umich.edu/bitstream/2027.42/145857/1/fnoohi_1.pd

    Features for matching people in different views

    No full text
    There have been significant advances in the computer vision field during the last decade. During this period, many methods have been developed that have been successful in solving challenging problems including Face Detection, Object Recognition and 3D Scene Reconstruction. The solutions developed by computer vision researchers have been widely adopted and used in many real-life applications such as those faced in the medical and security industry. Among the different branches of computer vision, Object Recognition has been an area that has advanced rapidly in recent years. The successful introduction of approaches such as feature extraction and description has been an important factor in the growth of this area. In recent years, researchers have attempted to use these approaches and apply them to other problems such as Content Based Image Retrieval and Tracking. In this work, we present a novel system that finds correspondences between people seen in different images. Unlike other approaches that rely on a video stream to track the movement of people between images, here we present a feature-based approach where we locate a target’s new location in an image, based only on its visual appearance. Our proposed system comprises three steps. In the first step, a set of features is extracted from the target’s appearance. A novel algorithm is developed that allows extraction of features from a target that is particularly suitable to the modelling task. In the second step, each feature is characterised using a combined colour and texture descriptor. Inclusion of information relating to both colour and texture of a feature add to the descriptor’s distinctiveness. Finally, the target’s appearance and pose is modelled as a collection of such features and descriptors. This collection is then used as a template that allows us to search for a similar combination of features in other images that correspond to the target’s new location. We have demonstrated the effectiveness of our system in locating a target’s new position in an image, despite differences in viewpoint, scale or elapsed time between the images. The characterisation of a target as a collection of features also allows our system to robustly deal with the partial occlusion of the target

    High-Level Facade Image Interpretation using Marked Point Processes

    Get PDF
    In this thesis, we address facade image interpretation as one essential ingredient for the generation of high-detailed, semantic meaningful, three-dimensional city-models. Given a single rectified facade image, we detect relevant facade objects such as windows, entrances, and balconies, which yield a description of the image in terms of accurate position and size of these objects. Urban digital three-dimensional reconstruction and documentation is an active area of research with several potential applications, e.g., in the area of digital mapping for navigation, urban planning, emergency management, disaster control or the entertainment industry. A detailed building model which is not just a geometric object enriched with texture, allows for semantic requests as the number of floors or the location of balconies and entrances. Facade image interpretation is one essential step in order to yield such models. In this thesis, we propose the interpretation of facade images by combining evidence for the occurrence of individual object classes which we derive from data, and prior knowledge which guides the image interpretation in its entirety. We present a three-step procedure which generates features that are suited to describe relevant objects, learns a representation that is suited for object detection, and that enables the image interpretation using the results of object detection while incorporating prior knowledge about typical configurations of facade objects, which we learn from training data. According to these three sub-tasks, our major achievements are: We propose a novel method for facade image interpretation based on a marked point process. Therefor, we develop a model for the description of typical configurations of facade objects and propose an image interpretation system which combines evidence derived from data and prior knowledge about typical configurations of facade objects. In order to generate evidence from data, we propose a feature type which we call shapelets. They are scale invariant and provide large distinctiveness for facade objects. Segments of lines, arcs, and ellipses serve as basic features for the generation of shapelets. Therefor, we propose a novel line simplification approach which approximates given pixel-chains by a sequence of lines, circular, and elliptical arcs. Among others, it is based on an adaption to Douglas-Peucker's algorithm, which is based on circles as basic geometric elements We evaluate each step separately. We show the effects of polyline segmentation and simplification on several images with comparable good or even better results, referring to a state-of-the-art algorithm, which proves their large distinctiveness for facade objects. Using shapelets we provide a reasonable classification performance on a challenging dataset, including intra-class variations, clutter, and scale changes. Finally, we show promising results for the facade interpretation system on several datasets and provide a qualitative evaluation which demonstrates the capability of complete and accurate detection of facade objectsHigh-Level Interpretation von Fassaden-Bildern unter Benutzung von Markierten PunktprozessenDas Thema dieser Arbeit ist die Interpretation von Fassadenbildern als wesentlicher Beitrag zur Erstellung hoch detaillierter, semantisch reichhaltiger dreidimensionaler Stadtmodelle. In rektifizierten Einzelaufnahmen von Fassaden detektieren wir relevante Objekte wie Fenster, Türen und Balkone, um daraus eine Bildinterpretation in Form von präzisen Positionen und Größen dieser Objekte abzuleiten. Die digitale dreidimensionale Rekonstruktion urbaner Regionen ist ein aktives Forschungsfeld mit zahlreichen Anwendungen, beispielsweise der Herstellung digitaler Kartenwerke für Navigation, Stadtplanung, Notfallmanagement, Katastrophenschutz oder die Unterhaltungsindustrie. Detaillierte Gebäudemodelle, die nicht nur als geometrische Objekte repräsentiert und durch eine geeignete Textur visuell ansprechend dargestellt werden, erlauben semantische Anfragen, wie beispielsweise nach der Anzahl der Geschosse oder der Position der Balkone oder Eingänge. Die semantische Interpretation von Fassadenbildern ist ein wesentlicher Schritt für die Erzeugung solcher Modelle. In der vorliegenden Arbeit lösen wir diese Aufgabe, indem wir aus Daten abgeleitete Evidenz für das Vorkommen einzelner Objekte mit Vorwissen kombinieren, das die Analyse der gesamten Bildinterpretation steuert. Wir präsentieren dafür ein dreistufiges Verfahren: Wir erzeugen Bildmerkmale, die für die Beschreibung der relevanten Objekte geeignet sind. Wir lernen, auf Basis abgeleiteter Merkmale, eine Repräsentation dieser Objekte. Schließlich realisieren wir die Bildinterpretation basierend auf der zuvor gelernten Repräsentation und dem Vorwissen über typische Konfigurationen von Fassadenobjekten, welches wir aus Trainingsdaten ableiten. Wir leisten dazu die folgenden wissenschaftlichen Beiträge: Wir schlagen eine neuartige Me-thode zur Interpretation von Fassadenbildern vor, die einen sogenannten markierten Punktprozess verwendet. Dafür entwickeln wir ein Modell zur Beschreibung typischer Konfigurationen von Fassadenobjekten und entwickeln ein Bildinterpretationssystem, welches aus Daten abgeleitete Evidenz und a priori Wissen über typische Fassadenkonfigurationen kombiniert. Für die Erzeugung der Evidenz stellen wir Merkmale vor, die wir Shapelets nennen und die skaleninvariant und durch eine ausgesprochene Distinktivität im Bezug auf Fassadenobjekte gekennzeichnet sind. Als Basismerkmale für die Erzeugung der Shapelets dienen Linien-, Kreis- und Ellipsensegmente. Dafür stellen wir eine neuartige Methode zur Vereinfachung von Liniensegmenten vor, die eine Pixelkette durch eine Sequenz von geraden Linienstücken und elliptischen Bogensegmenten approximiert. Diese basiert unter anderem auf einer Adaption des Douglas-Peucker Algorithmus, die anstelle gerader Linienstücke, Bogensegmente als geometrische Basiselemente verwendet. Wir evaluieren jeden dieser drei Teilschritte separat. Wir zeigen Ergebnisse der Liniensegmen-tierung anhand verschiedener Bilder und weisen dabei vergleichbare und teilweise verbesserte Ergebnisse im Vergleich zu bestehende Verfahren nach. Für die vorgeschlagenen Shapelets weisen wir in der Evaluation ihre diskriminativen Eigenschaften im Bezug auf Fassadenobjekte nach. Wir erzeugen auf einem anspruchsvollen Datensatz von skalenvariablen Fassadenobjekten, mit starker Variabilität der Erscheinung innerhalb der Klassen, vielversprechende Klassifikationsergebnisse, die die Verwendbarkeit der gelernten Shapelets für die weitere Interpretation belegen. Schließlich zeigen wir Ergebnisse der Interpretation der Fassadenstruktur anhand verschiedener Datensätze. Die qualitative Evaluation demonstriert die Fähigkeit des vorgeschlagenen Lösungsansatzes zur vollständigen und präzisen Detektion der genannten Fassadenobjekte

    Vision-Based Object Recognition and 3-D Pose Estimation Using Conic Features

    Get PDF
    This thesis deals with monocular vision-based object recognition and 3-D pose estimation based on conic features. Conic features including circles and ellipses are frequently observed in many man-made objects in real word as well as have the merit of robustness potentially in feature extraction in vision-based applications. Although the 3-D pose estimation problem of conic features in 3-D space has been studied well since 1990, the previous work has not provided a unique solution completely for full 3-D pose parameters (i.e., 3-orientations and 3-positions) due to complexity from high nonlinearity of a general conic. This thesis, therefore, renews conic features in a new perspective on geometric invariants in both 3-D space and 2-D projective space, incorporating other geometric features with conics. First, as the most essential step in dealing with conics, this thesis shows that the pose parameters of a circular feature in 3-D space can be derived analytically from incorporating a coplanar point. A procedure of pose parameter recovery is described in detail, and its performance is evaluated and discussed in view of pose estimation errors and sensitivity. Second, it is also revealed that the pose of an elliptic feature can be resolved when two coplanar points are incorporated on the basis of the polarity of two points for a conic in 2-D projective space. This thesis proposes a series of algorithms to determine the 3-D pose parameters uniquely, and evaluates the proposed method through a measure of estimation performance and sensitivity depending on point locations. Third, a pair of two conics is dealt with, which is regarded as an extension of the idea of the incorporation scheme to another conic feature from point features. Under the polarity concept, this thesis proves that the problem involving a pair of two conics can be formulated with the problem of one ellipse with two points so that its solution is derived in the same form as in the ellipse case. In order to treat two or more conic objects as well as to deal with an object recognition problem, the rest of thesis concentrates on the theoretical foundation of multiple object recognition. First, some effective modeling approaches are described. A general object model is specially designed to model multiple objects for object recognition and pose recovery in view of spatial geometry. In particular, this thesis defines a pairwise conic model that can describes the geometrical relation between two conics invariantly in 2-D projective space, which consists of a pairwise conic (PC), a pairwise conic invariant (PCI), and a pairwise conic pole (PCP). Based on the two kinds of models, an object learning and recognition system is proposed as a general framework for multiple object recognition. Considering simplicity and flexibility in object learning stage, this thesis introduces a semi-automatic learning scheme to construct the multiple object model from a model image at once. To utilize geometric relations among multiple objects effectively in object recognition, this thesis specifies some feature functions based on the pairwise conic model, and then describes an object recognition method in a fashion of linear-chain conditional random field (CRF). In particular, as a post refinement step of the recognition, a geometric alignment procedure is also proposed in algorithmic details to improve recognition performance against noisy conditions. Last, the multiple object recognition method is evaluated intensively through two practical applications that deal with a place recognition and an elevator button recognition problem for service robots. A series of experiment results supports the effectiveness of the proposed method, maintaining reliable performance against noisy conditions in the presence of perspective distortion and partial object occlusions.Contents Abstract . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . i Contents . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . iv List of Tables . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . viii List of Figures . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . ix 1 Introduction 1 1.1 Background . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1 1.2 Observations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3 1.3 Research objective and expected contribution . . . . . . . . . . . . . . . . . . 6 1.4 Organization of thesis . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8 2 3-D Pose Estimation of a Circular Feature 10 2.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10 2.1.1 Background . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10 2.1.2 Problem formulation . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11 2.1.3 Related work . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12 2.1.4 Notations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14 2.2 Preliminaries: an elliptic cone in 3-D space and its homogeneous representation in 2-D projective space . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15 2.2.1 Homogeneous representation . . . . . . . . . . . . . . . . . . . . . . . 15 2.2.2 Principal planes of a cone versus diagonalization of a conic matrix Q . 16 2.3 3-D interpretation of a circular feature for 3-D pose estimation . . . . . . . . 19 2.3.1 3-D orientation estimation . . . . . . . . . . . . . . . . . . . . . . . . . 19 2.3.2 3-D position estimation . . . . . . . . . . . . . . . . . . . . . . . . . . 22 2.3.3 Composition of homogeneous transformation and discrimination for the unique solution . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 23 2.4 Experiment results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 26 2.4.1 A numerical example . . . . . . . . . . . . . . . . . . . . . . . . . . . . 26 2.4.2 Evaluation of pose estimation performance . . . . . . . . . . . . . . . 29 2.5 Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 33 3 3-D Pose Estimation of an Elliptic Feature 35 3.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 35 3.1.1 Background . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 35 3.1.2 Problem statement . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 36 3.1.3 Related work . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 37 3.2 Interpretation of an elliptic feature with coplanar points in 2-D projective space 38 3.2.1 The minimal number of points for pose estimation . . . . . . . . . . . 39 3.2.2 Analysis of possible constraints for relative positions of two points to an ellipse . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 41 3.2.3 Feature selection scheme for stable homography estimation . . . . . . 43 3.3 3-D pose estimation algorithm . . . . . . . . . . . . . . . . . . . . . . . . . . . 46 3.3.1 Extraction of triangular features from an elliptic object . . . . . . . . 47 3.3.2 Homography decomposition . . . . . . . . . . . . . . . . . . . . . . . . 50 3.3.3 Composition of homogeneous transformation matrix with unique solution . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 52 3.4 Experiment results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 53 3.4.1 Experimental setup . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 53 3.4.2 Evaluation of the proposed method . . . . . . . . . . . . . . . . . . . . 54 3.5 Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 60 4 3-D Pose Estimation of a Pair of Conic Features 61 4.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 61 4.2 3-D pose estimation of a conic feature incorporated with line features . . . . 61 4.3 3-D pose estimation of a conic feature incorporated with another conic feature 63 4.3.1 Some examples of self-polar triangle and invariants . . . . . . . . . . . 65 4.3.2 3-D pose estimation of a pair of coplanar conics . . . . . . . . . . . . . 67 4.3.3 Examples of 3-D pose estimation of a conic feature incorporated with another conic feature . . . . . . . . . . . . . . . . . . . . . . . . . . . . 72 4.4 Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 76 5 Multiple Object Recognition Based on Pairwise Conic Model 77 5.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 77 5.2 Learning of geometric relation of multiple objects . . . . . . . . . . . . . . . . 78 5.3 Pairwise conic model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 80 5.3.1 De_nitions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 80 5.4 Multiple object recognition based on pairwise conic model and conditional random _elds . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 86 5.4.1 Graphical model for multiple object recognition . . . . . . . . . . . . . 86 5.4.2 Linear-chain conditional random _eld . . . . . . . . . . . . . . . . . . 87 5.4.3 Determination of low-level feature functions for multiple object recognition . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 90 5.4.4 Range selection trick for e_ciently computing the costs of low-level feature functions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 92 5.4.5 Evaluation of observation sequence . . . . . . . . . . . . . . . . . . . . 93 5.4.6 Object recognition based on hierarchical CRF . . . . . . . . . . . . . . 95 5.5 Geometric alignment algorithm . . . . . . . . . . . . . . . . . . . . . . . . . . 97 5.6 Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 103 6 Application to Place Recognition for Service Robots 105 6.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 105 6.1.1 Background . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 105 6.1.2 Problem statement . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 106 6.2 Feature extraction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 107 6.2.1 Detection of 2-D geometric shapes . . . . . . . . . . . . . . . . . . . . 107 6.2.2 Examples of shape feature extraction . . . . . . . . . . . . . . . . . . . 109 6.3 Object modeling . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 112 6.3.1 A place model that describes multiple landmark objects . . . . . . . . 112 6.3.2 Pairwise conic model . . . . . . . . . . . . . . . . . . . . . . . . . . . . 113 6.3.3 Incorporation of non-conic features with a pairwise conic model . . . . 114 6.4 Place learning and recognition system . . . . . . . . . . . . . . . . . . . . . . 121 6.4.1 HCRF-based recognition . . . . . . . . . . . . . . . . . . . . . . . . . . 122 6.5 Experiment results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 123 6.5.1 Experimental setup . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 123 6.5.2 Performance evaluation . . . . . . . . . . . . . . . . . . . . . . . . . . 127 6.6 Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 135 7 Application to Elevator Button Recognition 136 7.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 136 7.1.1 Background . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 136 7.1.2 Problem statement . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 138 7.1.3 Related work . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 138 7.2 Object modeling . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 140 7.2.1 Geometric model for multiple button objects . . . . . . . . . . . . . . 140 7.2.2 Pairwise conic model . . . . . . . . . . . . . . . . . . . . . . . . . . . . 141 7.3 Learning and recognition system . . . . . . . . . . . . . . . . . . . . . . . . . 141 7.3.1 Button object learning . . . . . . . . . . . . . . . . . . . . . . . . . . . 142 7.3.2 CRF-based recognition . . . . . . . . . . . . . . . . . . . . . . . . . . . 143 7.4 Experiment results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 147 7.4.1 Experimental setup . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 147 7.4.2 Performance evaluation . . . . . . . . . . . . . . . . . . . . . . . . . . 151 7.5 Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 158 8 Concluding remarks 159 8.1 Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 159 8.2 Further work . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 160 References 161 Summary (in Korean) 16

    Geometric and photometric affine invariant image registration

    Get PDF
    This thesis aims to present a solution to the correspondence problem for the registration of wide-baseline images taken from uncalibrated cameras. We propose an affine invariant descriptor that combines the geometry and photometry of the scene to find correspondences between both views. The geometric affine invariant component of the descriptor is based on the affine arc-length metric, whereas the photometry is analysed by invariant colour moments. A graph structure represents the spatial distribution of the primitive features; i.e. nodes correspond to detected high-curvature points, whereas arcs represent connectivities by extracted contours. After matching, we refine the search for correspondences by using a maximum likelihood robust algorithm. We have evaluated the system over synthetic and real data. The method is endemic to propagation of errors introduced by approximations in the system.BAE SystemsSelex Sensors and Airborne System
    corecore