14,059 research outputs found
Improving automation standards via semantic modelling: Application to ISA88
Standardization is essential for automation. Extensibility, scalability, and reusability are important features for automation software that rely in the efficient modelling of the addressed systems. The work presented here is from the ongoing development of a methodology for semi-automatic ontology construction methodology from technical documents. The main aim of this work is to systematically check the consistency of technical documents and support the improvement of technical document consistency. The formalization of conceptual models and the subsequent writing of technical standards are simultaneously analyzed, and guidelines proposed for application to future technical standards. Three paradigms are discussed for the development of domain ontologies from technical documents, starting from the current state of the art, continuing with the intermediate method presented and used in this paper, and ending with the suggested paradigm for the future. The ISA88 Standard is taken as a representative case study. Linguistic techniques from the semi-automatic ontology construction methodology is applied to the ISA88 Standard and different modelling and standardization aspects that are worth sharing with the automation community is addressed. This study discusses different paradigms for developing and sharing conceptual models for the subsequent development of automation software, along with presenting the systematic consistency checking methodPeer ReviewedPostprint (author's final draft
Operation Frames and Clubs in Kidney Exchange
A kidney exchange is a centrally-administered barter market where patients
swap their willing yet incompatible donors. Modern kidney exchanges use
2-cycles, 3-cycles, and chains initiated by non-directed donors (altruists who
are willing to give a kidney to anyone) as the means for swapping.
We propose significant generalizations to kidney exchange. We allow more than
one donor to donate in exchange for their desired patient receiving a kidney.
We also allow for the possibility of a donor willing to donate if any of a
number of patients receive kidneys. Furthermore, we combine these notions and
generalize them. The generalization is to exchange among organ clubs, where a
club is willing to donate organs outside the club if and only if the club
receives organs from outside the club according to given specifications. We
prove that unlike in the standard model, the uncapped clearing problem is
NP-complete.
We also present the notion of operation frames that can be used to sequence
the operations across batches, and present integer programming formulations for
the market clearing problems for these new types of organ exchanges.
Experiments show that in the single-donation setting, operation frames
improve planning by 34%--51%. Allowing up to two donors to donate in exchange
for one kidney donated to their designated patient yields a further increase in
social welfare.Comment: Published at IJCAI-1
Benchmarking Distributed Stream Data Processing Systems
The need for scalable and efficient stream analysis has led to the
development of many open-source streaming data processing systems (SDPSs) with
highly diverging capabilities and performance characteristics. While first
initiatives try to compare the systems for simple workloads, there is a clear
gap of detailed analyses of the systems' performance characteristics. In this
paper, we propose a framework for benchmarking distributed stream processing
engines. We use our suite to evaluate the performance of three widely used
SDPSs in detail, namely Apache Storm, Apache Spark, and Apache Flink. Our
evaluation focuses in particular on measuring the throughput and latency of
windowed operations, which are the basic type of operations in stream
analytics. For this benchmark, we design workloads based on real-life,
industrial use-cases inspired by the online gaming industry. The contribution
of our work is threefold. First, we give a definition of latency and throughput
for stateful operators. Second, we carefully separate the system under test and
driver, in order to correctly represent the open world model of typical stream
processing deployments and can, therefore, measure system performance under
realistic conditions. Third, we build the first benchmarking framework to
define and test the sustainable performance of streaming systems.
Our detailed evaluation highlights the individual characteristics and
use-cases of each system.Comment: Published at ICDE 201
Generative Adversarial Estimation of Channel Covariance in Vehicular Millimeter Wave Systems
Enabling highly-mobile millimeter wave (mmWave) systems is challenging
because of the huge training overhead associated with acquiring the channel
knowledge or designing the narrow beams. Current mmWave beam training and
channel estimation techniques do not normally make use of the prior beam
training or channel estimation observations. Intuitively, though, the channel
matrices are functions of the various elements of the environment. Learning
these functions can dramatically reduce the training overhead needed to obtain
the channel knowledge. In this paper, a novel solution that exploits machine
learning tools, namely conditional generative adversarial networks (GAN), is
developed to learn these functions between the environment and the channel
covariance matrices. More specifically, the proposed machine learning model
treats the covariance matrices as 2D images and learns the mapping function
relating the uplink received pilots, which act as RF signatures of the
environment, and these images. Simulation results show that the developed
strategy efficiently predicts the covariance matrices of the large-dimensional
mmWave channels with negligible training overhead.Comment: to appear in Asilomar Conference on Signals, Systems, and Computers,
Oct. 201
Large-scale associations between the leukocyte transcriptome and BOLD responses to speech differ in autism early language outcome subtypes.
Heterogeneity in early language development in autism spectrum disorder (ASD) is clinically important and may reflect neurobiologically distinct subtypes. Here, we identified a large-scale association between multiple coordinated blood leukocyte gene coexpression modules and the multivariate functional neuroimaging (fMRI) response to speech. Gene coexpression modules associated with the multivariate fMRI response to speech were different for all pairwise comparisons between typically developing toddlers and toddlers with ASD and poor versus good early language outcome. Associated coexpression modules were enriched in genes that are broadly expressed in the brain and many other tissues. These coexpression modules were also enriched in ASD-associated, prenatal, human-specific, and language-relevant genes. This work highlights distinctive neurobiology in ASD subtypes with different early language outcomes that is present well before such outcomes are known. Associations between neuroimaging measures and gene expression levels in blood leukocytes may offer a unique in vivo window into identifying brain-relevant molecular mechanisms in ASD
A method for encoding clinical datasets with SNOMED CT
<p>Abstract</p> <p>Background</p> <p>Over the past decade there has been a growing body of literature on how the Systematised Nomenclature of Medicine Clinical Terms (SNOMED CT) can be implemented and used in different clinical settings. Yet, for those charged with incorporating SNOMED CT into their organisation's clinical applications and vocabulary systems, there are few detailed encoding instructions and examples available to show how this can be done and the issues involved. This paper describes a heuristic method that can be used to encode clinical terms in SNOMED CT and an illustration of how it was applied to encode an existing palliative care dataset.</p> <p>Methods</p> <p>The encoding process involves: identifying input data items; cleaning the data items; encoding the cleaned data items; and exporting the encoded terms as output term sets. Four outputs are produced: the SNOMED CT reference set; interface terminology set; SNOMED CT extension set and unencodeable term set.</p> <p>Results</p> <p>The original palliative care database contained 211 data elements, 145 coded values and 37,248 free text values. We were able to encode ~84% of the terms, another ~8% require further encoding and verification while terms that had a frequency of fewer than five were not encoded (~7%).</p> <p>Conclusions</p> <p>From the pilot, it would seem our SNOMED CT encoding method has the potential to become a general purpose terminology encoding approach that can be used in different clinical systems.</p
ARTSCENE: A Neural System for Natural Scene Classification
How do humans rapidly recognize a scene? How can neural models capture this biological competence to achieve state-of-the-art scene classification? The ARTSCENE neural system classifies natural scene photographs by using multiple spatial scales to efficiently accumulate evidence for gist and texture. ARTSCENE embodies a coarse-to-fine Texture Size Ranking Principle whereby spatial attention processes multiple scales of scenic information, ranging from global gist to local properties of textures. The model can incrementally learn and predict scene identity by gist information alone and can improve performance through selective attention to scenic textures of progressively smaller size. ARTSCENE discriminates 4 landscape scene categories (coast, forest, mountain and countryside) with up to 91.58% correct on a test set, outperforms alternative models in the literature which use biologically implausible computations, and outperforms component systems that use either gist or texture information alone. Model simulations also show that adjacent textures form higher-order features that are also informative for scene recognition.National Science Foundation (NSF SBE-0354378); Office of Naval Research (N00014-01-1-0624
- …