14,059 research outputs found

    Improving automation standards via semantic modelling: Application to ISA88

    Get PDF
    Standardization is essential for automation. Extensibility, scalability, and reusability are important features for automation software that rely in the efficient modelling of the addressed systems. The work presented here is from the ongoing development of a methodology for semi-automatic ontology construction methodology from technical documents. The main aim of this work is to systematically check the consistency of technical documents and support the improvement of technical document consistency. The formalization of conceptual models and the subsequent writing of technical standards are simultaneously analyzed, and guidelines proposed for application to future technical standards. Three paradigms are discussed for the development of domain ontologies from technical documents, starting from the current state of the art, continuing with the intermediate method presented and used in this paper, and ending with the suggested paradigm for the future. The ISA88 Standard is taken as a representative case study. Linguistic techniques from the semi-automatic ontology construction methodology is applied to the ISA88 Standard and different modelling and standardization aspects that are worth sharing with the automation community is addressed. This study discusses different paradigms for developing and sharing conceptual models for the subsequent development of automation software, along with presenting the systematic consistency checking methodPeer ReviewedPostprint (author's final draft

    Operation Frames and Clubs in Kidney Exchange

    Full text link
    A kidney exchange is a centrally-administered barter market where patients swap their willing yet incompatible donors. Modern kidney exchanges use 2-cycles, 3-cycles, and chains initiated by non-directed donors (altruists who are willing to give a kidney to anyone) as the means for swapping. We propose significant generalizations to kidney exchange. We allow more than one donor to donate in exchange for their desired patient receiving a kidney. We also allow for the possibility of a donor willing to donate if any of a number of patients receive kidneys. Furthermore, we combine these notions and generalize them. The generalization is to exchange among organ clubs, where a club is willing to donate organs outside the club if and only if the club receives organs from outside the club according to given specifications. We prove that unlike in the standard model, the uncapped clearing problem is NP-complete. We also present the notion of operation frames that can be used to sequence the operations across batches, and present integer programming formulations for the market clearing problems for these new types of organ exchanges. Experiments show that in the single-donation setting, operation frames improve planning by 34%--51%. Allowing up to two donors to donate in exchange for one kidney donated to their designated patient yields a further increase in social welfare.Comment: Published at IJCAI-1

    Benchmarking Distributed Stream Data Processing Systems

    Full text link
    The need for scalable and efficient stream analysis has led to the development of many open-source streaming data processing systems (SDPSs) with highly diverging capabilities and performance characteristics. While first initiatives try to compare the systems for simple workloads, there is a clear gap of detailed analyses of the systems' performance characteristics. In this paper, we propose a framework for benchmarking distributed stream processing engines. We use our suite to evaluate the performance of three widely used SDPSs in detail, namely Apache Storm, Apache Spark, and Apache Flink. Our evaluation focuses in particular on measuring the throughput and latency of windowed operations, which are the basic type of operations in stream analytics. For this benchmark, we design workloads based on real-life, industrial use-cases inspired by the online gaming industry. The contribution of our work is threefold. First, we give a definition of latency and throughput for stateful operators. Second, we carefully separate the system under test and driver, in order to correctly represent the open world model of typical stream processing deployments and can, therefore, measure system performance under realistic conditions. Third, we build the first benchmarking framework to define and test the sustainable performance of streaming systems. Our detailed evaluation highlights the individual characteristics and use-cases of each system.Comment: Published at ICDE 201

    Generative Adversarial Estimation of Channel Covariance in Vehicular Millimeter Wave Systems

    Full text link
    Enabling highly-mobile millimeter wave (mmWave) systems is challenging because of the huge training overhead associated with acquiring the channel knowledge or designing the narrow beams. Current mmWave beam training and channel estimation techniques do not normally make use of the prior beam training or channel estimation observations. Intuitively, though, the channel matrices are functions of the various elements of the environment. Learning these functions can dramatically reduce the training overhead needed to obtain the channel knowledge. In this paper, a novel solution that exploits machine learning tools, namely conditional generative adversarial networks (GAN), is developed to learn these functions between the environment and the channel covariance matrices. More specifically, the proposed machine learning model treats the covariance matrices as 2D images and learns the mapping function relating the uplink received pilots, which act as RF signatures of the environment, and these images. Simulation results show that the developed strategy efficiently predicts the covariance matrices of the large-dimensional mmWave channels with negligible training overhead.Comment: to appear in Asilomar Conference on Signals, Systems, and Computers, Oct. 201

    Large-scale associations between the leukocyte transcriptome and BOLD responses to speech differ in autism early language outcome subtypes.

    Get PDF
    Heterogeneity in early language development in autism spectrum disorder (ASD) is clinically important and may reflect neurobiologically distinct subtypes. Here, we identified a large-scale association between multiple coordinated blood leukocyte gene coexpression modules and the multivariate functional neuroimaging (fMRI) response to speech. Gene coexpression modules associated with the multivariate fMRI response to speech were different for all pairwise comparisons between typically developing toddlers and toddlers with ASD and poor versus good early language outcome. Associated coexpression modules were enriched in genes that are broadly expressed in the brain and many other tissues. These coexpression modules were also enriched in ASD-associated, prenatal, human-specific, and language-relevant genes. This work highlights distinctive neurobiology in ASD subtypes with different early language outcomes that is present well before such outcomes are known. Associations between neuroimaging measures and gene expression levels in blood leukocytes may offer a unique in vivo window into identifying brain-relevant molecular mechanisms in ASD

    A method for encoding clinical datasets with SNOMED CT

    Get PDF
    <p>Abstract</p> <p>Background</p> <p>Over the past decade there has been a growing body of literature on how the Systematised Nomenclature of Medicine Clinical Terms (SNOMED CT) can be implemented and used in different clinical settings. Yet, for those charged with incorporating SNOMED CT into their organisation's clinical applications and vocabulary systems, there are few detailed encoding instructions and examples available to show how this can be done and the issues involved. This paper describes a heuristic method that can be used to encode clinical terms in SNOMED CT and an illustration of how it was applied to encode an existing palliative care dataset.</p> <p>Methods</p> <p>The encoding process involves: identifying input data items; cleaning the data items; encoding the cleaned data items; and exporting the encoded terms as output term sets. Four outputs are produced: the SNOMED CT reference set; interface terminology set; SNOMED CT extension set and unencodeable term set.</p> <p>Results</p> <p>The original palliative care database contained 211 data elements, 145 coded values and 37,248 free text values. We were able to encode ~84% of the terms, another ~8% require further encoding and verification while terms that had a frequency of fewer than five were not encoded (~7%).</p> <p>Conclusions</p> <p>From the pilot, it would seem our SNOMED CT encoding method has the potential to become a general purpose terminology encoding approach that can be used in different clinical systems.</p

    ARTSCENE: A Neural System for Natural Scene Classification

    Full text link
    How do humans rapidly recognize a scene? How can neural models capture this biological competence to achieve state-of-the-art scene classification? The ARTSCENE neural system classifies natural scene photographs by using multiple spatial scales to efficiently accumulate evidence for gist and texture. ARTSCENE embodies a coarse-to-fine Texture Size Ranking Principle whereby spatial attention processes multiple scales of scenic information, ranging from global gist to local properties of textures. The model can incrementally learn and predict scene identity by gist information alone and can improve performance through selective attention to scenic textures of progressively smaller size. ARTSCENE discriminates 4 landscape scene categories (coast, forest, mountain and countryside) with up to 91.58% correct on a test set, outperforms alternative models in the literature which use biologically implausible computations, and outperforms component systems that use either gist or texture information alone. Model simulations also show that adjacent textures form higher-order features that are also informative for scene recognition.National Science Foundation (NSF SBE-0354378); Office of Naval Research (N00014-01-1-0624
    • …
    corecore