225 research outputs found

    Attention-based distributed speech enhancement for unconstrained microphone arrays with varying number of nodes

    Full text link
    Speech enhancement promises higher efficiency in ad-hoc microphone arrays than in constrained microphone arrays thanks to the wide spatial coverage of the devices in the acoustic scene. However, speech enhancement in ad-hoc microphone arrays still raises many challenges. In particular, the algorithms should be able to handle a variable number of microphones, as some devices in the array might appear or disappear. In this paper, we propose a solution that can efficiently process the spatial information captured by the different devices of the microphone array, while being robust to a link failure. To do this, we use an attention mechanism in order to put more weight on the relevant signals sent throughout the array and to neglect the redundant or empty channels

    A Trimodel SAR Semisupervised Recognition Method Based on Attention-Augmented Convolutional Networks

    Get PDF
    Semisupervised learning in synthetic aperture radars (SARs) is one of the research hotspots in the field of radar image automatic target recognition. It can efficiently deal with challenging environments where there are insufficient labeled samples and large unlabeled samples in the SAR dataset. In recent years, consistency regularization methods in semisupervised learning have shown considerable improvement in recognition accuracy and efficiency. Current consistency regularization approaches suffer from two main shortcomings: first, extracting all of the relevant information in the image target is difficult owing to the inability of conventional convolutional neural networks to capture global relational information; second, the standard teacher–student regularization methodology causes confirmation biases due to the high coupling between teacher and student models. This article adopts an innovative trimodel semisupervised method based on attention-augmented convolutional networks to address the aforementioned obstacles. Specifically, we develop an attention mechanism incorporating a novel positional embedding method based on recurrent neural networks and integrate this with a standard convolutional network as a feature extractor, to improve the network's ability to extract global feature information from images. Furthermore, we address the confirmation bias problem by introducing a classmate model to the standard teacher–student structure and utilize the model to impose a weak consistency constraint designed on the student to weaken the strong coupling between the teacher and the student. Comparative experiments on the Moving and Stationary Target Acquisition and Recognition dataset show that our method outperforms state-of-the-art semisupervised methods in terms of recognition accuracy, demonstrating its potential as a new benchmark approach for the deep learning and SAR research community

    Electronics for Sensors

    Get PDF
    The aim of this Special Issue is to explore new advanced solutions in electronic systems and interfaces to be employed in sensors, describing best practices, implementations, and applications. The selected papers in particular concern photomultiplier tubes (PMTs) and silicon photomultipliers (SiPMs) interfaces and applications, techniques for monitoring radiation levels, electronics for biomedical applications, design and applications of time-to-digital converters, interfaces for image sensors, and general-purpose theory and topologies for electronic interfaces

    Integrated GANs: Semi-Supervised SAR Target Recognition

    Get PDF
    With the advantage of working in all weathers and all days, synthetic aperture radar (SAR) imaging systems have a great application value. As an efficient image generation and recognition model, generative adversarial networks (GANs) have been applied to SAR image analysis and achieved promising performance. However, the cost of labeling a large number of SAR images limits the performance of the developed approaches and aggravates the mode collapsing problem. This paper presents a novel approach namely Integrated GANs (I-GAN), which consists of a conditional GANs, an unconditional GANs and a classifier, to achieve semi-supervised generation and recognition simultaneously. The unconditional GANs assist the conditional GANs to increase the diversity of the generated images. A co-training method for the conditional GANs and the classifier is proposed to enrich the training samples. Since our model is capable of representing training images with rich characteristics, the classifier can achieve better recognition accuracy. Experiments on the Moving and Stationary Target Acquisition and Recognition (MSTAR) dataset proves that our method achieves better results in accuracy when labeled samples are insufficient, compared against other state-of-the-art techniques

    On the use of deep learning for phase recovery

    Full text link
    Phase recovery (PR) refers to calculating the phase of the light field from its intensity measurements. As exemplified from quantitative phase imaging and coherent diffraction imaging to adaptive optics, PR is essential for reconstructing the refractive index distribution or topography of an object and correcting the aberration of an imaging system. In recent years, deep learning (DL), often implemented through deep neural networks, has provided unprecedented support for computational imaging, leading to more efficient solutions for various PR problems. In this review, we first briefly introduce conventional methods for PR. Then, we review how DL provides support for PR from the following three stages, namely, pre-processing, in-processing, and post-processing. We also review how DL is used in phase image processing. Finally, we summarize the work in DL for PR and outlook on how to better use DL to improve the reliability and efficiency in PR. Furthermore, we present a live-updating resource (https://github.com/kqwang/phase-recovery) for readers to learn more about PR.Comment: 82 pages, 32 figure

    Complex land cover classifications and physical properties retrieval of tropical forests using multi-source remote sensing

    Get PDF
    The work presented in this thesis mainly focuses on two subjects related to the application of remote sensing data: (1) for land cover classification combining optical sensor, texture features generated from spectral information and synthetic aperture radar (SAR) features, and (2) to develop a non-destructive approach for above ground biomass (AGB) and forest attributes estimation employing multi-source remote sensing data (i.e. optical data, SAR backscatter) combined with in-situ data. Information provided by reliable land cover map is useful for management of forest resources to support sustainable forest management, whereas the generation of the non-destructive approach to model forest biophysical properties (e.g. AGB and stem volume) is required to assess the forest resources more efficiently and cost-effective, and coupled with remote sensing data the model can be applied over large forest areas. This work considers study sites over tropical rain forest landscape in Indonesia characterized by different successional stages and complex vegetation structure including tropical peatland forests. The thesis begins with a brief introduction and the state of the art explaining recent trends on monitoring and modeling of forest resources using remote sensing data and approach. The research works on the integration of spectral information and texture features for forest cover mapping is presented subsequently, followed by development of a non-destructive approach for AGB and forest parameters predictions and modeling. Ultimately, this work evaluates the potential of mosaic SAR data for AGB modeling and the fusion of optical and SAR data for peatlands discrimination. The results show that the inclusion of geostatistics texture features improved the classification accuracy of optical Landsat ETM data. Moreover, the fusion of SAR and optical data enhanced the peatlands discrimination over tropical peat swamp forest. For forest stand parameters modeling, neural networks method resulted in lower error estimate than standard multi-linear regression technique, and the combination of non-destructive measurement (i.e. stem number) and remote sensing data improved the model accuracy. The up scaling of stem volume and biomass estimates using Kriging method and bi-temporal ETM image also provide favorable estimate results upon comparison with the land cover map.Die in dieser Dissertation präsentierten Ergebnisse konzentrieren sich hauptsächlich auf zwei Themen mit Bezug zur angewandten Fernerkundung: 1) Der Klassifizierung von Oberflächenbedeckung basierend auf der Verknüpfung von optischen Sensoren, Textureigenschaften erzeugt durch Spektraldaten und Synthetic-Aperture-Radar (SAR) features und 2) die Entwicklung eines nichtdestruktiven Verfahrens zur Bestimmung oberirdischer Biomasse (AGB) und weiterer Waldeigenschaften mittels multi-source Fernerkundungsdaten (optische Daten, SAR Rückstreuung) sowie in-situ Daten. Eine zuverlässige Karte der Landbedeckung dient der Unterstützung von nachhaltigem Waldmanagement, während eine nichtdestruktive Herangehensweise zur Modellierung von biophysikalischen Waldeigenschaften (z.B. AGB und Stammvolumen) für eine effiziente und kostengünstige Beurteilung der Waldressourcen notwendig ist. Durch die Kopplung mit Fernerkundungsdaten kann das Modell auf große Waldflächen übertragen werden. Die vorliegende Arbeit berücksichtigt Untersuchungsgebiete im tropischen Regenwald Indonesiens, welche durch verschiedene Regenerations- und Sukzessionsstadien sowie komplexe Vegetationsstrukturen, inklusive tropischer Torfwälder, gekennzeichnet sind. Am Anfang der Arbeit werden in einer kurzen Einleitung der Stand der Forschung und die neuesten Forschungstrends in der Überwachung und Modellierung von Waldressourcen mithilfe von Fernerkundungsdaten dargestellt. Anschließend werden die Forschungsergebnisse der Kombination von Spektraleigenschaften und Textureigenschaften zur Waldbedeckungskartierung erläutert. Desweiteren folgen Ergebnisse zur Entwicklung eines nichtdestruktiven Ansatzes zur Vorhersage und Modellierung von AGB und Waldeigenschaften, zur Auswertung von Mosaik- SAR Daten für die Modellierung von AGB, sowie zur Fusion optischer mit SAR Daten für die Identifizierung von Torfwäldern. Die Ergebnisse zeigen, dass die Einbeziehung von geostatistischen Textureigenschaften die Genauigkeit der Klassifikation von optischen Landsat ETM Daten gesteigert hat. Desweiteren führte die Fusion von SAR und optischen Daten zu einer Verbesserung der Unterscheidung zwischen Torfwäldern und tropischen Sumpfwäldern. Bei der Modellierung der Waldparameter führte die Neural-Network-Methode zu niedrigeren Fehlerschätzungen als die multiple Regressions. Die Kombination von nichtdestruktiven Messungen (z.B. Stammzahl) und Fernerkundungsdaten führte zu einer Steigerung der Modellgenauigkeit. Die Hochskalierung des Stammvolumens und Schätzungen der Biomasse mithilfe von Kriging und bi-temporalen ETM Daten lieferten positive Schätzergebnisse im Vergleich zur Landbedeckungskarte

    Deep neural network generation for image classification within resource-constrained environments using evolutionary and hand-crafted processes

    Get PDF
    Constructing Convolutional Neural Networks (CNN) models is a manual process requiringexpert knowledge and trial and error. Background research highlights the following knowledge gaps. 1) existing efficiency-focused CNN models make design choices that impact model performance. Better ways are needed to construct accurate models for resourceconstrained environments that lack graphics processing units (GPU’s) to speed up model inference time such as CCTV cameras and IoT devices. 2) Existing methods for automatically designing CNN architectures do not explore the search space effectively for the best solution and 3) existing methods for automatically designing CNN architectures do not exploit modern model architecture design patterns such as residual connections. The lack of residual connections means the model depth is limited owing to the vanishing gradient problem. Furthermore, existing methods for automatically designing CNN architectures adopt search strategies that make them vulnerable to local minima traps. Better techniques to construct efficient CNN models, and automated approaches that can produce accurate deep model constructions advance many areas such as hazard detection, medical diagnosis and robotics in both academia and industry. The work undertaken during this research are 1) the proposal of an efficient and accurate CNN architecture for resource-constrained environments owing to a novel block structure containing 1x3 and 3x1 convolutions to save computational cost, 2) proposed a particle swarm optimization (PSO) method of automatically constructing efficient deep CNN architectures with greater accuracy by proposing a novel encoding and search strategy, 3) proposed a PSO based method of automatically constructing deeper CNN models with improved accuracy by proposing a novel encoding scheme that employs residual connections, in novel search mechanism that follows the global and neighbouring best leaders. The main findings of this research are 1) the proposed efficiency-focused CNN model outperformed MobileNetV2 by 13.43% in respect to accuracy, and 39.63% in respect to efficiency, measured in floating-point operations. A reduction in floating-point operations means the model has the potential for faster inference times which is beneficial to applications within resource-constrained environments without GPU’s such as CCTV cameras. 2) the proposed automatic CNN generation technique outperformed existing methods by 7.58% in respect to accuracy and a 63% improvement in search time efficiency owing to the proposal of more efficient architectures speeding up the search process and 3) the proposed automatic deep residual CNN generation method improved model accuracy by 4.43% when compared against related studies owing to deeper model construction and improvements in the search process. The proposed search process embeds human knowledge of constructing deep residual networks and provides constraint settings which can be used to limit the proposed models depth and width. The ability to constrain a models depth and width is important as it ensures the upper bounds of a proposed model will fit within the constraints of resource-constrained environments

    Machine Learning in Sensors and Imaging

    Get PDF
    Machine learning is extending its applications in various fields, such as image processing, the Internet of Things, user interface, big data, manufacturing, management, etc. As data are required to build machine learning networks, sensors are one of the most important technologies. In addition, machine learning networks can contribute to the improvement in sensor performance and the creation of new sensor applications. This Special Issue addresses all types of machine learning applications related to sensors and imaging. It covers computer vision-based control, activity recognition, fuzzy label classification, failure classification, motor temperature estimation, the camera calibration of intelligent vehicles, error detection, color prior model, compressive sensing, wildfire risk assessment, shelf auditing, forest-growing stem volume estimation, road management, image denoising, and touchscreens

    Optical and radar remotely sensed data for large-area wildlife habitat mapping

    Get PDF
    Wildlife habitat mapping strongly supports applications in natural resource management, environmental conservation, impacts of anthropogenic activity, perturbed ecosystem restoration, species-at-risk recovery and species inventory. Remote sensing has long been identified as a feasible and effective technology for large-area wildlife habitat mapping. However, existing and future uncertainties in remote sensing will definitely have a significant effect on relevant scientific research, such as the limitation of Landsat-series data; the negative impact of cloud and cloud shadows (CCS) in optical imagery; and landscape pattern analysis using remote sensing classification products. This thesis adopted a manuscript-style format; it addresses these challenges (or uncertainties) and opportunities through exploring the state-of-the-art optical and radar remotely sensed data for large-area wildlife habitat mapping, and investigating their feasibility and applicability primarily by comparison either on the level of direct remote sensing products (e.g. classification accuracy) or indirect ecological model (e.g. presence/absence and frequency of use model based on landscape pattern analysis). A framework designed to identify and investigate the potential remotely sensed data, including Disaster Monitoring Constellation (DMC), Landsat Thematic Mapper (TM), Indian Remote Sensing (IRS), and RADARSAT-2, has been developed. The chosen DMC and RADARSAT-2 imagery have acceptable capability of addressing the existing and potential challenges (or uncertainties) in remote sensing of large-area habitat mapping, in order to produce cloud-free thematic maps for the study of wildlife habitat. A quantitative comparison between Landsat-based and IRS-based analyses showed that the characteristics of remote sensing products play an important role in landscape pattern analysis to build grizzly bear presence/absence and frequency of use models
    • …
    corecore