86 research outputs found

    Design Of Computer Vision Systems For Optimizing The Threat Detection Accuracy

    Get PDF
    This dissertation considers computer vision (CV) systems in which a central monitoring station receives and analyzes the video streams captured and delivered wirelessly by multiple cameras. It addresses how the bandwidth can be allocated to various cameras by presenting a cross-layer solution that optimizes the overall detection or recognition accuracy. The dissertation presents and develops a real CV system and subsequently provides a detailed experimental analysis of cross-layer optimization. Other unique features of the developed solution include employing the popular HTTP streaming approach, utilizing homogeneous cameras as well as heterogeneous ones with varying capabilities and limitations, and including a new algorithm for estimating the effective medium airtime. The results show that the proposed solution significantly improves the CV accuracy. Additionally, the dissertation features an improved neural network system for object detection. The proposed system considers inherent video characteristics and employs different motion detection and clustering algorithms to focus on the areas of importance in consecutive frames, allowing the system to dynamically and efficiently distribute the detection task among multiple deployments of object detection neural networks. Our experimental results indicate that our proposed method can enhance the mAP (mean average precision), execution time, and required data transmissions to object detection networks. Finally, as recognizing an activity provides significant automation prospects in CV systems, the dissertation presents an efficient activity-detection recurrent neural network that utilizes fast pose/limbs estimation approaches. By combining object detection with pose estimation, the domain of activity detection is shifted from a volume of RGB (Red, Green, and Blue) pixel values to a time-series of relatively small one-dimensional arrays, thereby allowing the activity detection system to take advantage of highly capable neural networks that have been trained on large GPU clusters for thousands of hours. Consequently, capable activity detection systems with considerably fewer training sets and processing hours can be built

    Enhanced clustering analysis pipeline for performance analysis of parallel applications

    Get PDF
    Clustering analysis is widely used to stratify data in the same cluster when they are similar according to the specific metrics. We can use the cluster analysis to group the CPU burst of a parallel application, and the regions on each process in-between communication calls or calls to the parallel runtime. The resulting clusters obtained are the different computational trends or phases that appear in the application. These clusters are useful to understand the behavior of the computation part of the application and focus the analyses on those that present performance issues. Although density-based clustering algorithms are a powerful and efficient tool to summarize this type of information, their traditional user-guided clustering methodology has many shortcomings and deficiencies in dealing with the complexity of data, the diversity of data structures, high-dimensionality of data, and the dramatic increase in the amount of data. Consequently, the majority of DBSCAN-like algorithms have weaknesses to handle high-dimensionality and/or Multi-density data, and they are sensitive to their hyper-parameter configuration. Furthermore, extracting insight from the obtained clusters is an intuitive and manual task. To mitigate these weaknesses, we have proposed a new unified approach to replace the user-guided clustering with an automated clustering analysis pipeline, called Enhanced Cluster Identification and Interpretation (ECII) pipeline. To build the pipeline, we propose novel techniques including Robust Independent Feature Selection, Feature Space Curvature Map, Organization Component Analysis, and hyper-parameters tuning to feature selection, density homogenization, cluster interpretation, and model selection which are the main components of our machine learning pipeline. This thesis contributes four new techniques to the Machine Learning field with a particular use case in Performance Analytics field. The first contribution is a novel unsupervised approach for feature selection on noisy data, called Robust Independent Feature Selection (RIFS). Specifically, we choose a feature subset that contains most of the underlying information, using the same criteria as the Independent component analysis. Simultaneously, the noise is separated as an independent component. The second contribution of the thesis is a parametric multilinear transformation method to homogenize cluster densities while preserving the topological structure of the dataset, called Feature Space Curvature Map (FSCM). We present a new Gravitational Self-organizing Map to model the feature space curvature by plugging the concepts of gravity and fabric of space into the Self-organizing Map algorithm to mathematically describe the density structure of the data. To homogenize the cluster density, we introduce a novel mapping mechanism to project the data from the non-Euclidean curved space to a new Euclidean flat space. The third contribution is a novel topological-based method to study potentially complex high-dimensional categorized data by quantifying their shapes and extracting fine-grain insights from them to interpret the clustering result. We introduce our Organization Component Analysis (OCA) method for the automatic arbitrary cluster-shape study without an assumption about the data distribution. Finally, to tune the DBSCAN hyper-parameters, we propose a new tuning mechanism by combining techniques from machine learning and optimization domains, and we embed it in the ECII pipeline. Using this cluster analysis pipeline with the CPU burst data of a parallel application, we provide the developer/analyst with a high-quality SPMD computation structure detection with the added value that reflects the fine grain of the computation regions.El análisis de conglomerados se usa ampliamente para estratificar datos en el mismo conglomerado cuando son similares según las métricas específicas. Nosotros puede usar el análisis de clúster para agrupar la ráfaga de CPU de una aplicación paralela y las regiones en cada proceso intermedio llamadas de comunicación o llamadas al tiempo de ejecución paralelo. Los clusters resultantes obtenidos son las diferentes tendencias computacionales o fases que aparecen en la solicitud. Estos clusters son útiles para entender el comportamiento de la parte de computación del aplicación y centrar los análisis en aquellos que presenten problemas de rendimiento. Aunque los algoritmos de agrupamiento basados en la densidad son una herramienta poderosa y eficiente para resumir este tipo de información, su La metodología tradicional de agrupación en clústeres guiada por el usuario tiene muchas deficiencias y deficiencias al tratar con la complejidad de los datos, la diversidad de estructuras de datos, la alta dimensionalidad de los datos y el aumento dramático en la cantidad de datos. En consecuencia, el La mayoría de los algoritmos similares a DBSCAN tienen debilidades para manejar datos de alta dimensionalidad y/o densidad múltiple, y son sensibles a su configuración de hiperparámetros. Además, extraer información de los clústeres obtenidos es una forma intuitiva y tarea manual Para mitigar estas debilidades, hemos propuesto un nuevo enfoque unificado para reemplazar el agrupamiento guiado por el usuario con un canalización de análisis de agrupamiento automatizado, llamada canalización de identificación e interpretación de clúster mejorada (ECII). para construir el tubería, proponemos técnicas novedosas que incluyen la selección robusta de características independientes, el mapa de curvatura del espacio de características, Análisis de componentes de la organización y ajuste de hiperparámetros para la selección de características, homogeneización de densidad, agrupación interpretación y selección de modelos, que son los componentes principales de nuestra canalización de aprendizaje automático. Esta tesis aporta cuatro nuevas técnicas al campo de Machine Learning con un caso de uso particular en el campo de Performance Analytics. La primera contribución es un enfoque novedoso no supervisado para la selección de características en datos ruidosos, llamado Robust Independent Feature. Selección (RIFS).Específicamente, elegimos un subconjunto de funciones que contiene la mayor parte de la información subyacente, utilizando el mismo criterios como el análisis de componentes independientes. Simultáneamente, el ruido se separa como un componente independiente. La segunda contribución de la tesis es un método de transformación multilineal paramétrica para homogeneizar densidades de clústeres mientras preservando la estructura topológica del conjunto de datos, llamado Mapa de Curvatura del Espacio de Características (FSCM). Presentamos un nuevo Gravitacional Mapa autoorganizado para modelar la curvatura del espacio característico conectando los conceptos de gravedad y estructura del espacio en el Algoritmo de mapa autoorganizado para describir matemáticamente la estructura de densidad de los datos. Para homogeneizar la densidad del racimo, introducimos un mecanismo de mapeo novedoso para proyectar los datos del espacio curvo no euclidiano a un nuevo plano euclidiano espacio. La tercera contribución es un nuevo método basado en topología para estudiar datos categorizados de alta dimensión potencialmente complejos mediante cuantificando sus formas y extrayendo información detallada de ellas para interpretar el resultado de la agrupación. presentamos nuestro Método de análisis de componentes de organización (OCA) para el estudio automático de forma arbitraria de conglomerados sin una suposición sobre el distribución de datos.Postprint (published version

    Structural Performance Comparison of Parallel Software Applications

    Get PDF
    With rising complexity of high performance computing systems and their parallel software, performance analysis and optimization has become essential in the development of efficient applications. The comparison of performance data is a key operation required in performance analysis. An analyst may conduct different types of comparisons in order to understand the performance properties of an application. One use case is comparing performance data from multiple measurements. Typical examples for such comparisons are before/after comparisons when applying optimizations or changing code versions. Besides comparing performance between multiple runs, also comparing performance characteristics across the parallel execution streams of an application is essential to detect performance problems. This is typically useful to detect imbalances, outliers, or changing runtime behavior during the execution of an application. While such comparisons are straightforward for the aggregated data in performance profiles, only limited solutions exist for comparing event traces. Trace-based analysis, i.e., the collection of fine-grained information on individual application events with timestamps and application context, has proven to be a powerful technique. The detailed performance information included in event traces make them very suitable for performance analysis. However, this level of detail also presents a challenge because it implies a large and overwhelming amount of data. Currently, users need to perform manual comparison of event traces, which is extremely challenging and time consuming because of the large volume of detailed data and the need to correctly line up trace events. To fill the gap of missing solutions for automatic comparison of event traces, this work proposes a set of techniques that automatically align traces. The alignment allows their structural comparison and the highlighting of differences between them. A set of novel metrics provide the user with an objective measure of the differences between traces, both in terms of differences in the event stream and timing differences across events. An additional important aspect of trace-based analysis is the visualization of performance data in event timelines. This has proven to be a powerful approach for the detection of various types of performance problems. However, visualization of large numbers of event timelines quickly hits the limits of available display resolution. Likewise, identifying performance problems is challenging in the large amount of visualized performance data. To alleviate these problems this work proposes two new approaches for event timeline visualization. First, novel folding strategies for event timelines facilitate visual scalability and provide powerful overviews of performance data at the same time. Second, this work presents an effective approach that automatically identifies and highlights several types of performance critical sections in an application run. This approach identifies time dominant functions of an application and subsequently uses them to analyze runtime imbalances throughout the application run. Intuitive visualizations present the resulting runtime variations and guide the analyst to performance hot spots. Evaluations with benchmarks and real-world applications assess all introduced techniques. The effectiveness of the comparison approaches is demonstrated by showing automatically detected performance issues and structural differences between different versions of applications and across parallel execution streams. Case studies showcase the capabilities of the event timeline visualization techniques by demonstrating scalable performance data visualizations and detecting performance problems and code inefficiencies in real-world applications

    A Review of Indoor Millimeter Wave Device-based Localization and Device-free Sensing Technologies and Applications

    Full text link
    The commercial availability of low-cost millimeter wave (mmWave) communication and radar devices is starting to improve the penetration of such technologies in consumer markets, paving the way for large-scale and dense deployments in fifth-generation (5G)-and-beyond as well as 6G networks. At the same time, pervasive mmWave access will enable device localization and device-free sensing with unprecedented accuracy, especially with respect to sub-6 GHz commercial-grade devices. This paper surveys the state of the art in device-based localization and device-free sensing using mmWave communication and radar devices, with a focus on indoor deployments. We first overview key concepts about mmWave signal propagation and system design. Then, we provide a detailed account of approaches and algorithms for localization and sensing enabled by mmWaves. We consider several dimensions in our analysis, including the main objectives, techniques, and performance of each work, whether each research reached some degree of implementation, and which hardware platforms were used for this purpose. We conclude by discussing that better algorithms for consumer-grade devices, data fusion methods for dense deployments, as well as an educated application of machine learning methods are promising, relevant and timely research directions.Comment: 43 pages, 13 figures. Accepted in IEEE Communications Surveys & Tutorials (IEEE COMST

    Risk-Aware Planning for Sensor Data Collection

    Get PDF
    With the emergence of low-cost unmanned air vehicles, civilian and military organizations are quickly identifying new applications for affordable, large-scale collectives to support and augment human efforts via sensor data collection. In order to be viable, these collectives must be resilient to the risk and uncertainty of operating in real-world environments. Previous work in multi-agent planning has avoided planning for the loss of agents in environments with risk. In contrast, this dissertation presents a problem formulation that includes the risk of losing agents, the effect of those losses on the mission being executed, and provides anticipatory planning algorithms that consider risk. We conduct a thorough analysis of the effects of risk on path-based planning, motivating new solution methods. We then use hierarchical clustering to generate risk-aware plans for a variable number of agents, outperforming traditional planning methods. Next, we provide a mechanism for distributed negotiation of stable plans, utilizing coalitional game theory to provide cost allocation methods that we prove to be fair and stable. Centralized planning with redundancy is then explored, planning for parallel task completion to mitigate risk and provide further increased expected value. Finally, we explore the role of cost uncertainty as additional source of risk, using bi-objective optimization to generate sets of alternative plans. We demonstrate the capability of our algorithms on randomly generated problem instances, showing an improvement over traditional multi-agent planning methods as high as 500% on very large problem instances

    Prognostic and health management of critical aircraft systems and components: an overview

    Get PDF
    This article belongs to the Special Issue Feature Papers in Fault Diagnosis & Sensors 2023Prognostic and health management (PHM) plays a vital role in ensuring the safety and reliability of aircraft systems. The process entails the proactive surveillance and evaluation of the state and functional effectiveness of crucial subsystems. The principal aim of PHM is to predict the remaining useful life (RUL) of subsystems and proactively mitigate future breakdowns in order to minimize consequences. The achievement of this objective is helped by employing predictive modeling techniques and doing real-time data analysis. The incorporation of prognostic methodologies is of utmost importance in the execution of condition-based maintenance (CBM), a strategic approach that emphasizes the prioritization of repairing components that have experienced quantifiable damage. Multiple methodologies are employed to support the advancement of prognostics for aviation systems, encompassing physics-based modeling, data-driven techniques, and hybrid prognosis. These methodologies enable the prediction and mitigation of failures by identifying relevant health indicators. Despite the promising outcomes in the aviation sector pertaining to the implementation of PHM, there exists a deficiency in the research concerning the efficient integration of hybrid PHM applications. The primary aim of this paper is to provide a thorough analysis of the current state of research advancements in prognostics for aircraft systems, with a specific focus on prominent algorithms and their practical applications and challenges. The paper concludes by providing a detailed analysis of prospective directions for future research within the field.European Union funding: 95568

    Heuristics approaches for three-dimensional strip packing and multiple carrier transportation plans

    Get PDF
    In transport logistic operations, an efficient delivery plan and better utilisation of vehicles will result in fuel cost savings, reduced working hours and even reduction of carbon dioxide emissions. This thesis proposes various algorithmic approaches to generate improved performance in automated vehicle load packing and route planning. First, modifications to best-fit heuristic methodologies are proposed and then incorporated into a simple but effective “look-ahead” heuristic procedure. The results obtained are very competitive and in some cases best-known results are found for different sets of constraints on three-dimensional strip packing problems. Secondly, a review and comparison of different clustering techniques in transport route planning is presented. This study shows that the algorithmic approach performs according to the specific type of real-world transport route planning scenario under consideration. This study helps to achieve a better understanding of how to conduct the automated generation of vehicle routes that meet the specific conditions required in the operations of a transport logistics company. Finally, a new approach to measuring the quality of transportation route plans is presented showing how this procedure has a positive effect on the quality of the generated route plans. In summary, this thesis proposes new tailored and effective heuristic methodologies that have been tested and incorporated into the real-world operations of a transport logistics company. The research work presented here is a modest yet significant advance to better understanding and solving the difficult problems of vehicle loading and routing in real-world scenarios

    Dynamics of Hybrid Zones at a Continental Scale

    Get PDF
    Hybridization has traditionally been viewed as a happenstance that negatively impacts populations, but is now recognized as an important evolutionary mechanism that can substantially impact the evolutionary trajectories of gene pools, influence adaptive capacity, and contravene or reinforce divergence. Physiographic processes are important drivers of dispersal, alternately funneling populations into isolation, promoting divergence, or facilitating secondary contact of diverged populations, increasing the potential for hybridization. In North America, glacial-interglacial cycles and geomorphological changes have provided a dynamic backdrop over the last two million years that promoted such oscillations of population contraction and expansion. These biogeographic processes have resulted in regional hybrid zones where hybridization spans generations Herein, I explored hybrid zones in two species complexes of reptiles across Eastern, Central, and Southwestern North America. Hybrid zones can influence evolutionary trajectories, and understanding the mechanisms underlying their formation is important for defining appropriate management strategies and can help avoid actions that would inadvertently lead to new hybrid zones. Chapter I assessed differential introgression in a complex of terrestrial turtles, the American Box Turtles (Terrapene spp.), from a contemporary hybrid zone in the southeastern United States. Transcriptomic loci were correlated with environmental predictors to evaluate mechanisms engendering maladapted hybrids and adaptive introgression. Selection against hybrids predominated for inter-specifics but directional introgression did so in conspecifics. Outlier loci also primarily correlated with temperature, reflecting the temperature dependency of ectotherms and underscoring their vulnerability to climate change. Chapter II performed a robust assessment of recently developed machine learning (M-L) approaches to delimit four Terrapene species and evaluate the impact of data filtering and M-L parameter choices. Parameter selections were varied to determine their effects in resolving clusters. The results provide necessary recommendations on using M-L for species delimitation in species complexes defined by secondary contact. These data exemplify usage of M-L software in a phylogenetically complex group. Chapter III describes an R package to visualize some of the analyses from Chapter I. Current software to generate genomic clines does not include functions to visualize the results. Thus, I wrote an API (application programming interface) that does so and also performs other genomic and geographic cline-related tasks. Chapter IV examines historical and contemporary phylogeographic patterns in the Massasaugas (Sistrurus spp.), a type of dwarf rattlesnake found across the Southwest and Central Great Plains. In the Southwest, S. tergeminus tergeminus and S. t. edwardsii putatively diverged in the absence of strong physiographic barriers and physical glaciers, suggesting primary divergence. In contrast, a disjunct population of S. t. tergeminus in Missouri reflects potentially historical secondary contact with S. catenatus. These taxa represent contrasting examples of divergence resulting from alternative phylogeographic processes and contextualizes evolutionarily significant and management units. Combined, the four chapters present population genomic data to elucidate impacts of phylogeographic processes on hybrid zones at a continental scale. The data will promote effective conservation management strategies, as many species in the focal regions have been affected by anthropogenic pressures. In this sense, the results can be extrapolated to co-distributed taxa with similar phylogeographic histories
    corecore