614 research outputs found

    Aerial Image Analysis using Deep Learning for Electrical Overhead Line Network Asset Management

    Get PDF
    Electricity networks are critical infrastructure, delivering vital energy services. Due to the significant number, variety and distribution of electrical network overhead line assets, energy network operators spend millions annually on inspection and maintenance programmes. Currently, inspection involves acquiring and manually analysing aerial images. This is labour intensive and subjective. Along with costs associated with helicopter or drone operations, data analysis represents a significant financial burden to network operators. We propose an approach to automating assessment of the condition of electrical towers. Importantly, we train machine learning tower classifiers without using condition labels for individual components of interest. Instead, learning is supervised using only condition labels for towers in their entirety. This enables us to use a real-world industry dataset without needing costly additional human labelling of thousands of individual components. Our prototype first detects instances of components in multiple images of each tower, using Mask R-CNN or RetinaNet. It then predicts tower condition ratings using one of two approaches: (i) component instance classifiers trained using class labels transferred from towers to each of their detected component instances, or (ii) multiple instance learning classifiers based on bags of detected instances. Instance or bag class predictions are aggregated to obtain tower condition ratings. Evaluation used a dataset with representative tower images and associated condition ratings covering a range of component types, scenes, environmental conditions, and viewpoints. We report experiments investigating classification of towers based on the condition of their multiple insulator and U-bolt components. Insulators and their U-bolts were detected with average precision of 96.7 and 97.9, respectively. Tower classification achieved areas under ROC curves of 0.94 and 0.98 for insulator condition and U-bolt condition ratings, respectively. Thus we demonstrate that tower condition classifiers can be trained effectively without labelling the condition of individual components

    A new method to analyse mosaics based on Symmetry Group theory applied to Islamic Geometric Patterns

    Get PDF
    [EN] This article presents a new method for analysing mosaics based on the mathematical principles of Symmetry Groups. This method has been developed to get the understanding present in patterns by extracting the objects that form them, their lattice, and the Wallpaper Group. The main novelty of this method resides in the creation of a higher level of knowledge based on objects, which makes it possible to classify the objects, to extract their main features (Point Group, principal axes, etc.), and the relationships between them. In order to validate the method, several tests were carried out on a set of Islamic Geometric Patterns from different sources, for which the Wallpaper Group has been successfully obtained in 85% of the cases. This method can be applied to any kind of pattern that presents a Wallpaper Group. Possible applications of this computational method include pattern classification, cataloguing of ceramic coatings, creating databases of decorative patterns, creating pattern designs, pattern comparison between different cultures, tile cataloguing, and so on.The authors wish to thank the Patronato de la Alhambra y Generalife (Granada, Spain) and the Patronato del Real Alcázar de Sevilla (Seville, Spain) for their valuable collaboration in this research work.Albert Gil, FE.; Gomis Martí, JM.; Blasco, J.; Valiente González, JM.; Aleixos Borrás, MN. (2015). A new method to analyse mosaics based on Symmetry Group theory applied to Islamic Geometric Patterns. Computer Vision and Image Understanding. 130:54-70. doi:10.1016/j.cviu.2014.09.002S547013

    Learning scale-variant and scale-invariant features for deep image classification

    Get PDF
    Convolutional Neural Networks (CNNs) require large image corpora to be trained on classification tasks. The variation in image resolutions, sizes of objects and patterns depicted, and image scales, hampers CNN training and performance, because the task-relevant information varies over spatial scales. Previous work attempting to deal with such scale variations focused on encouraging scale-invariant CNN representations. However, scale-invariant representations are incomplete representations of images, because images contain scale-variant information as well. This paper addresses the combined development of scale-invariant and scale-variant representations. We propose a multi- scale CNN method to encourage the recognition of both types of features and evaluate it on a challenging image classification task involving task-relevant characteristics at multiple scales. The results show that our multi-scale CNN outperforms single-scale CNN. This leads to the conclusion that encouraging the combined development of a scale-invariant and scale-variant representation in CNNs is beneficial to image recognition performance

    Audio-Material Modeling and Reconstruction for Multimodal Interaction

    Get PDF
    Interactive virtual environments enable the creation of training simulations, games, and social applications. These virtual environments can create a sense of presence in the environment: a sensation that its user is truly in another location. To maintain presence, interactions with virtual objects should engage multiple senses. Furthermore, multisensory input should be consistent, e.g. a virtual bowl that visually appears plastic should also sound like plastic when dropped on the floor. In this dissertation, I propose methods to improve the perceptual realism of virtual object impact sounds and ensure consistency between those sounds and the input from other senses. Recreating the impact sound of a real-world object requires an accurate estimate of that object's material parameters. The material parameters that affect impact sound---collectively forming the audio-material---include the material damping parameters for a damping model. I propose and evaluate damping models and use them to estimate material damping parameters for real-world objects. I also consider how interaction with virtual objects can be made more consistent between the senses of sight, hearing, and touch. First, I present a method for modeling the damping behavior of impact sounds, using generalized proportional damping to both estimate more expressive material damping parameters from recorded impact sounds and perform impact sound synthesis. Next, I present a method for estimating material damping parameters in the presence of confounding factors and with no knowledge of the object's shape. To accomplish this, a probabilistic damping model captures various external effects to produce robust damping parameter estimates. Next, I present a method for consistent multimodal interaction with textured surfaces. Texture maps serve as a single unified representation of mesoscopic detail for the purposes of visual rendering, sound synthesis, and rigid-body simulation. Finally, I present a method for geometry and material classification using multimodal audio-visual input. Using this method, a real-world scene can be scanned and virtually reconstructed while accurately modeling both the visual appearances and audio-material parameters of each object.Doctor of Philosoph
    • …
    corecore