10,820 research outputs found

    Analysis of Professional Trajectories using Disconnected Self-Organizing Maps

    Full text link
    In this paper we address an important economic question. Is there, as mainstream economic theory asserts it, an homogeneous labor market with mechanisms which govern supply and demand for work, producing an equilibrium with its remarkable properties? Using the Panel Study of Income Dynamics (PSID) collected on the period 1984-2003, we study the situations of American workers with respect to employment. The data include all heads of household (men or women) as well as the partners who are on the labor market, working or not. They are extracted from the complete survey and we compute a few relevant features which characterize the worker's situations. To perform this analysis, we suggest using a Self-Organizing Map (SOM, Kohonen algorithm) with specific structure based on planar graphs, with disconnected components (called D-SOM), especially interesting for clustering. We compare the results to those obtained with a classical SOM grid and a star-shaped map (called SOS). Each component of D-SOM takes the form of a string and corresponds to an organized cluster. From this clustering, we study the trajectories of the individuals among the classes by using the transition probability matrices for each period and the corresponding stationary distributions. As a matter of fact, we find clear evidence of heterogeneous parts, each one with high homo-geneity, representing situations well identified in terms of activity and wage levels and in degree of stability in the workplace. These results and their interpretation in economic terms contribute to the debate about flexibility which is commonly seen as a way to obtain a better level of equilibrium on the labor market

    Nonlinear data driven techniques for process monitoring

    Get PDF
    The goal of this research is to develop process monitoring technology capable of taking advantage of the large stores of data accumulating in modern chemical plants. There is demand for new techniques for the monitoring of non-linear topology and behavior, and this research presents a topological preservation method for process monitoring using Self Organizing Maps (SOM). The novel architecture presented adapts SOM to a full spectrum of process monitoring tasks including fault detection, fault identification, fault diagnosis, and soft sensing. The key innovation of the new technique is its use of multiple SOM (MSOM) in the data modeling process as well as the use of a Gaussian Mixture Model (GMM) to model the probability density function of classes of data. For comparison, a linear process monitoring technique based on Principal Component Analysis (PCA) is also used to demonstrate the improvements SOM offers. Data for the computational experiments was generated using a simulation of the Tennessee Eastman process (TEP) created in Simulink by (Ricker 1996). Previous studies focus on step changes from normal operations, but this work adds operating regimes with time dependent dynamics not previously considered with a SOM. Results show that MSOM improves upon both linear PCA as well as the standard SOM technique using one map for fault diagnosis, and also shows a superior ability to isolate which variables in the data are responsible for the faulty condition. With respect to soft sensing, SOM and MSOM modeled the compositions equally well, showing that no information was lost in dividing the map representation of process data. Future research will attempt to validate the technique on a real chemical process

    Clustering and its Application in Requirements Engineering

    Get PDF
    Large scale software systems challenge almost every activity in the software development life-cycle, including tasks related to eliciting, analyzing, and specifying requirements. Fortunately many of these complexities can be addressed through clustering the requirements in order to create abstractions that are meaningful to human stakeholders. For example, the requirements elicitation process can be supported through dynamically clustering incoming stakeholders’ requests into themes. Cross-cutting concerns, which have a significant impact on the architectural design, can be identified through the use of fuzzy clustering techniques and metrics designed to detect when a theme cross-cuts the dominant decomposition of the system. Finally, traceability techniques, required in critical software projects by many regulatory bodies, can be automated and enhanced by the use of cluster-based information retrieval methods. Unfortunately, despite a significant body of work describing document clustering techniques, there is almost no prior work which directly addresses the challenges, constraints, and nuances of requirements clustering. As a result, the effectiveness of software engineering tools and processes that depend on requirements clustering is severely limited. This report directly addresses the problem of clustering requirements through surveying standard clustering techniques and discussing their application to the requirements clustering process

    Cluster analysis of the signal curves in perfusion DCE-MRI datasets

    Get PDF
    Pathological studies show that tumors consist of different sub-regions with more homogeneous vascular properties during their growth. In addition, destroying tumor's blood supply is the target of most cancer therapies. Finding the sub-regions in the tissue of interest with similar perfusion patterns provides us with valuable information about tissue structure and angiogenesis. This information on cancer therapy, for example, can be used in monitoring the response of the cancer treatment to the drug. Cluster analysis of perfusion curves assays to find sub-regions with a similar perfusion pattern. The present work focuses on the cluster analysis of perfusion curves, measured by dynamic contrast enhanced magnetic resonance imaging (DCE-MRI). The study, besides searching for the proper clustering method, follows two other major topics, the choice of an appropriate similarity measure, and determining the number of clusters. These three subjects are connected to each other in such a way that success in one direction will help solving the other problems. This work introduces a new similarity measure, parallelism measure (PM), for comparing the parallelism in the washout phase of the signal curves. Most of the previous works used the Euclidean distance as the measure of dissimilarity. However, the Euclidean distance does not take the patterns of the signal curves into account and therefore for comparing the signal curves is not sufficient. To combine the advantages of both measures a two-steps clustering is developed. The two-steps clustering uses two different similarity measures, the introduced PM measure and Euclidean distance in two consecutive steps. The results of two-steps clustering are compared with the results of other clustering methods. The two-steps clustering besides good performance has some other advantages. The granularity and the number of clusters are controlled by thresholds defined by considering the noise in signal curves. The method is easy to implement and is robust against noise. The focus of the work is mainly the cluster analysis of breast tumors in DCE-MRI datasets. The possibility to adopt the method for liver datasets is studied as well

    A cell outage management framework for dense heterogeneous networks

    Get PDF
    In this paper, we present a novel cell outage management (COM) framework for heterogeneous networks with split control and data planes-a candidate architecture for meeting future capacity, quality-of-service, and energy efficiency demands. In such an architecture, the control and data functionalities are not necessarily handled by the same node. The control base stations (BSs) manage the transmission of control information and user equipment (UE) mobility, whereas the data BSs handle UE data. An implication of this split architecture is that an outage to a BS in one plane has to be compensated by other BSs in the same plane. Our COM framework addresses this challenge by incorporating two distinct cell outage detection (COD) algorithms to cope with the idiosyncrasies of both data and control planes. The COD algorithm for control cells leverages the relatively larger number of UEs in the control cell to gather large-scale minimization-of-drive-test report data and detects an outage by applying machine learning and anomaly detection techniques. To improve outage detection accuracy, we also investigate and compare the performance of two anomaly-detecting algorithms, i.e., k-nearest-neighbor- and local-outlier-factor-based anomaly detectors, within the control COD. On the other hand, for data cell COD, we propose a heuristic Grey-prediction-based approach, which can work with the small number of UE in the data cell, by exploiting the fact that the control BS manages UE-data BS connectivity and by receiving a periodic update of the received signal reference power statistic between the UEs and data BSs in its coverage. The detection accuracy of the heuristic data COD algorithm is further improved by exploiting the Fourier series of the residual error that is inherent to a Grey prediction model. Our COM framework integrates these two COD algorithms with a cell outage compensation (COC) algorithm that can be applied to both planes. Our COC solution utilizes an actor-critic-based reinforcement learning algorithm, which optimizes the capacity and coverage of the identified outage zone in a plane, by adjusting the antenna gain and transmission power of the surrounding BSs in that plane. The simulation results show that the proposed framework can detect both data and control cell outage and compensate for the detected outage in a reliable manner

    Self-Organized Maps

    Get PDF
    Se han obtenido los siguientes resultados: (1) Estudio de topologías bidimensionales alternativas: se muestra la importancia de topologías alternativas basadas en áreas ajenas como las teselaciones. (2) Estudio comparativo de topologías en una, dos y tres dimensiones: se revela la influencia de la dimensión en el funcionamiento de una SOM a escala local y global. (3) Estudio de alternativas al movimiento euclídeo: se propone y presenta la alternativa FRSOM al algoritmo SOM clásico. En FRSOM, las neuronas esquivan barreras predefinidas en su movimiento. Las conclusiones más relevantes que emanan de esta Tesis Doctoral son las siguientes: (1) La calidad del clustering y de la preservación topológica de una SOM puede ser mejorada mediante el uso de topologías alternativas y también evitando regiones prohibidas que no contribuyan significativamente al Error Cuadrático Medio (ECM). (2) La dimensióon de la SOM que obtiene mejores resultados es la propia dimensión intrínseca de los datos. Además, en general, valores bajos para la dimensión de la SOM producen mejores resultados en términos del ECM, y valores altos ocasionan mejor aprendizaje de la estructura de los datos.Los mapas auto-organizados o redes de Kohonen (SOM por sus siglas en inglés, self-organizing map) fueron introducidos por el profesor finlandés Teuvo Kalevi Kohonen en los años 80. Un mapa auto-organizado es una herramienta que analiza datos en muchas dimensiones con relaciones complejas entre ellos y los reduce o representa en, usualmente, una, dos o tres dimensiones. La propiedad más importante de una SOM es que preserva las propiedades topológicas de los datos, es decir, que datos próximos aparecen próximos en la representación. La literatura relacionada con los mapas auto-organizados y sus aplicaciones es muy diversa y numerosa. Las neuronas en un mapa auto-organizado clásico están distribuidas en una topología (o malla) bidimensional cuadrada o hexagonal y las distancias entre ellas son distancias euclídeas. Una de las disciplinas de investigación en SOM consiste en la modificación y generalización del algoritmo SOM. Esta Tesis Doctoral por compendio de publicaciones se centra en esta línea de investigación. En concreto, los objetivos desarrollados han sido el estudio de topologías bidimensionales alternativas, el estudio comparativo de topologías de una, dos y tres dimensiones y el estudio de variaciones para la distancia y movimientos euclídeos. Estos objetivos se han abordado mediante el método científico a través de las siguientes fases: aprehensión de resultados conocidos, planteamiento de hipótesis, propuesta de métodos alternativos, confrontación de métodos mediante experimentación, aceptación y rechazo de las diversas hipótesis mediante métodos estadísticos

    Unsupervised model-based clustering for typological classification of Middle Bronze Age flanged axes

    Get PDF
    The classification of Western European flanged axes dating to the Middle Bronze Age (1650–1350 BC) is very complex. Many types of axe have been identified, some of which have numerous variant forms. In the current French terminology, all axes are divided into two generic groups: namely "Atlantic" (Atlantique) and "Eastern" (Orientale). Each of these generic groups, however, is highly polymorphic, so that it is often very difficult for the operator to classify individual axes with absolute confidence and certainty. In order to overcome such problems, a new shape classification is proposed, using morphometric analysis (Elliptic Fourier Analysis) followed by unsupervised model-based clustering and discriminant analysis, both based on Gaussian mixture modelling. Together, these methods produce a clearer pattern, which is independently validated by the spatial distribution of the findings, and multinomial scan statistics. This approach is fast, reproducible, and operator-independent, allowing artefacts of unknown membership to be classified rapidly. The method is designed to be amendable by the introduction of new artefacts, in the light of future discoveries. This method can be adapted to suit many other archaeological artefacts, providing information about the material, social and cultural relations of ancient populations

    A Neural Model of How Horizontal and Interlaminar Connections of Visual Cortex Develop into Adult Circuits that Carry Out Perceptual Grouping and Learning

    Full text link
    A neural model suggests how horizontal and interlaminar connections in visual cortical areas Vl and V2 develop within a laminar cortical architecture and give rise to adult visual percepts. The model suggests how mechanisms that control cortical development in the infant lead to properties of adult cortical anatomy, neurophysiology, and visual perception. The model clarifies how excitatory and inhibitory connections can develop stably by maintaining a balance between excitation and inhibition. The growth of long-range excitatory horizontal connections between layer 2/3 pyramidal cells is balanced against that of short-range disynaptic interneuronal connections. The growth of excitatory on-center connections from layer 6-to-4 is balanced against that of inhibitory interneuronal off-surround connections. These balanced connections interact via intracortical and intercortical feedback to realize properties of perceptual grouping, attention, and perceptual learning in the adult, and help to explain the observed variability in the number and temporal distribution of spikes emitted by cortical neurons. The model replicates cortical point spread functions and psychophysical data on the strength of real and illusory contours. The on-center off-surround layer 6-to-4 circuit enables top-clown attentional signals from area V2 to modulate, or attentionally prime, layer 4 cells in area Vl without fully activating them. This modulatory circuit also enables adult perceptual learning within cortical area Vl and V2 to proceed in a stable way.Defense Advanced Research Projects Agency and the Office of Naval Research (N00014-95-1-0409); National Science Foundation (IRI-97-20333); Office of Naval Research (N00014-95-1-0657

    Spatial temporal dynamics of neighborhood quality of life : Charlotte, NC

    Get PDF
    Quality of life (QoL) is an encompassing measure of a neighborhood's condition, describing the well-being an individual may expect by residing in a particular place. Over time, some or all of these conditions will change for the better or worse, yet the driving forces behind the dynamics of neighborhood-level QoL are not well understood. The purpose of this dissertation is to gain a better comprehension of the patterns, trajectories, and explanatory factors of change across the multidimensional QoL conditions of neighborhoods. Utilizing neighborhoods in Charlotte, NC over the course of the 2000-2010 decade as a case study, this dissertation employs three complementary analytical approaches to examine spatial, multidimensional dynamics: Markov Chains, self-organizing maps, and a set of cross-lagged panel models. Results highlight the role of spatial spillovers in shaping the change process; a neighborhood's mobility in terms of QoL is not independent of its immediate surroundings. Geographically, older, inner-ring suburban neighborhoods are shown to be most vulnerable to declines across multiple QoL dimensions; middle-age housing further proves to be a significant explanatory predictor of changes in crime concentrations, relative economic status, youth social indicators, and homeownership rates, thus supporting economic filtering theories of neighborhood change. Neighborhoods characterized by the highest QoL attributes are the most stable through time. Lower-income neighborhoods are found to be heterogeneous in terms of their corresponding social problems, and a temporal, reciprocal relationship between crime and youth-related problems is revealed. Improvements to the lowest QoL neighborhoods were heightened at the peak of the housing and economic boom in the city, but following the great recession, many of these neighborhoods reverted back to their conditions earlier in the decade, illuminating the shifting dynamics before, during, and after the recent, great recession. Policy implications of results are discussed
    corecore