79,381 research outputs found
Input variable selection in time-critical knowledge integration applications: A review, analysis, and recommendation paper
This is the post-print version of the final paper published in Advanced Engineering Informatics. The published article is available from the link below. Changes resulting from the publishing process, such as peer review, editing, corrections, structural formatting, and other quality control mechanisms may not be reflected in this document. Changes may have been made to this work since it was submitted for publication. Copyright @ 2013 Elsevier B.V.The purpose of this research is twofold: first, to undertake a thorough appraisal of existing Input Variable Selection (IVS) methods within the context of time-critical and computation resource-limited dimensionality reduction problems; second, to demonstrate improvements to, and the application of, a recently proposed time-critical sensitivity analysis method called EventTracker to an environment science industrial use-case, i.e., sub-surface drilling.
Producing time-critical accurate knowledge about the state of a system (effect) under computational and data acquisition (cause) constraints is a major challenge, especially if the knowledge required is critical to the system operation where the safety of operators or integrity of costly equipment is at stake. Understanding and interpreting, a chain of interrelated events, predicted or unpredicted, that may or may not result in a specific state of the system, is the core challenge of this research. The main objective is then to identify which set of input data signals has a significant impact on the set of system state information (i.e. output). Through a cause-effect analysis technique, the proposed technique supports the filtering of unsolicited data that can otherwise clog up the communication and computational capabilities of a standard supervisory control and data acquisition system.
The paper analyzes the performance of input variable selection techniques from a series of perspectives. It then expands the categorization and assessment of sensitivity analysis methods in a structured framework that takes into account the relationship between inputs and outputs, the nature of their time series, and the computational effort required. The outcome of this analysis is that established methods have a limited suitability for use by time-critical variable selection applications. By way of a geological drilling monitoring scenario, the suitability of the proposed EventTracker Sensitivity Analysis method for use in high volume and time critical input variable selection problems is demonstrated.E
The Five Factor Model of personality and evaluation of drug consumption risk
The problem of evaluating an individual's risk of drug consumption and misuse
is highly important. An online survey methodology was employed to collect data
including Big Five personality traits (NEO-FFI-R), impulsivity (BIS-11),
sensation seeking (ImpSS), and demographic information. The data set contained
information on the consumption of 18 central nervous system psychoactive drugs.
Correlation analysis demonstrated the existence of groups of drugs with
strongly correlated consumption patterns. Three correlation pleiades were
identified, named by the central drug in the pleiade: ecstasy, heroin, and
benzodiazepines pleiades. An exhaustive search was performed to select the most
effective subset of input features and data mining methods to classify users
and non-users for each drug and pleiad. A number of classification methods were
employed (decision tree, random forest, -nearest neighbors, linear
discriminant analysis, Gaussian mixture, probability density function
estimation, logistic regression and na{\"i}ve Bayes) and the most effective
classifier was selected for each drug. The quality of classification was
surprisingly high with sensitivity and specificity (evaluated by leave-one-out
cross-validation) being greater than 70\% for almost all classification tasks.
The best results with sensitivity and specificity being greater than 75\% were
achieved for cannabis, crack, ecstasy, legal highs, LSD, and volatile substance
abuse (VSA).Comment: Significantly extended report with 67 pages, 27 tables, 21 figure
Quantum Google in a Complex Network
We investigate the behavior of the recently proposed quantum Google
algorithm, or quantum PageRank, in large complex networks. Applying the quantum
algorithm to a part of the real World Wide Web, we find that the algorithm is
able to univocally reveal the underlying scale-free topology of the network and
to clearly identify and order the most relevant nodes (hubs) of the graph
according to their importance in the network structure. Moreover, our results
show that the quantum PageRank algorithm generically leads to changes in the
hierarchy of nodes. In addition, as compared to its classical counterpart, the
quantum algorithm is capable to clearly highlight the structure of secondary
hubs of the network, and to partially resolve the degeneracy in importance of
the low lying part of the list of rankings, which represents a typical
shortcoming of the classical PageRank algorithm. Complementary to this study,
our analysis shows that the algorithm is able to clearly distinguish scale-free
networks from other widespread and important classes of complex networks, such
as Erd\H{o}s-R\'enyi networks and hierarchical graphs. We show that the ranking
capabilities of the quantum PageRank algorithm are related to an increased
stability with respect to a variation of the damping parameter that
appears in the Google algorithm, and to a more clearly pronounced power-law
behavior in the distribution of importance among the nodes, as compared to the
classical algorithm. Finally, we study to which extent the increased
sensitivity of the quantum algorithm persists under coordinated attacks of the
most important nodes in scale-free and Erd\H{o}s-R\'enyi random graphs
CASSL: Curriculum Accelerated Self-Supervised Learning
Recent self-supervised learning approaches focus on using a few thousand data
points to learn policies for high-level, low-dimensional action spaces.
However, scaling this framework for high-dimensional control require either
scaling up the data collection efforts or using a clever sampling strategy for
training. We present a novel approach - Curriculum Accelerated Self-Supervised
Learning (CASSL) - to train policies that map visual information to high-level,
higher- dimensional action spaces. CASSL orders the sampling of training data
based on control dimensions: the learning and sampling are focused on few
control parameters before other parameters. The right curriculum for learning
is suggested by variance-based global sensitivity analysis of the control
space. We apply our CASSL framework to learning how to grasp using an adaptive,
underactuated multi-fingered gripper, a challenging system to control. Our
experimental results indicate that CASSL provides significant improvement and
generalization compared to baseline methods such as staged curriculum learning
(8% increase) and complete end-to-end learning with random exploration (14%
improvement) tested on a set of novel objects
Knowledge-aware Complementary Product Representation Learning
Learning product representations that reflect complementary relationship
plays a central role in e-commerce recommender system. In the absence of the
product relationships graph, which existing methods rely on, there is a need to
detect the complementary relationships directly from noisy and sparse customer
purchase activities. Furthermore, unlike simple relationships such as
similarity, complementariness is asymmetric and non-transitive. Standard usage
of representation learning emphasizes on only one set of embedding, which is
problematic for modelling such properties of complementariness. We propose
using knowledge-aware learning with dual product embedding to solve the above
challenges. We encode contextual knowledge into product representation by
multi-task learning, to alleviate the sparsity issue. By explicitly modelling
with user bias terms, we separate the noise of customer-specific preferences
from the complementariness. Furthermore, we adopt the dual embedding framework
to capture the intrinsic properties of complementariness and provide geometric
interpretation motivated by the classic separating hyperplane theory. Finally,
we propose a Bayesian network structure that unifies all the components, which
also concludes several popular models as special cases. The proposed method
compares favourably to state-of-art methods, in downstream classification and
recommendation tasks. We also develop an implementation that scales efficiently
to a dataset with millions of items and customers
- âŠ