60,087 research outputs found
Probabilistic Matching: Causal Inference under Measurement Errors
The abundance of data produced daily from large variety of sources has
boosted the need of novel approaches on causal inference analysis from
observational data. Observational data often contain noisy or missing entries.
Moreover, causal inference studies may require unobserved high-level
information which needs to be inferred from other observed attributes. In such
cases, inaccuracies of the applied inference methods will result in noisy
outputs. In this study, we propose a novel approach for causal inference when
one or more key variables are noisy. Our method utilizes the knowledge about
the uncertainty of the real values of key variables in order to reduce the bias
induced by noisy measurements. We evaluate our approach in comparison with
existing methods both on simulated and real scenarios and we demonstrate that
our method reduces the bias and avoids false causal inference conclusions in
most cases.Comment: In Proceedings of International Joint Conference Of Neural Networks
(IJCNN) 201
OpenTag: Open Attribute Value Extraction from Product Profiles [Deep Learning, Active Learning, Named Entity Recognition]
Extraction of missing attribute values is to find values describing an
attribute of interest from a free text input. Most past related work on
extraction of missing attribute values work with a closed world assumption with
the possible set of values known beforehand, or use dictionaries of values and
hand-crafted features. How can we discover new attribute values that we have
never seen before? Can we do this with limited human annotation or supervision?
We study this problem in the context of product catalogs that often have
missing values for many attributes of interest.
In this work, we leverage product profile information such as titles and
descriptions to discover missing values of product attributes. We develop a
novel deep tagging model OpenTag for this extraction problem with the following
contributions: (1) we formalize the problem as a sequence tagging task, and
propose a joint model exploiting recurrent neural networks (specifically,
bidirectional LSTM) to capture context and semantics, and Conditional Random
Fields (CRF) to enforce tagging consistency, (2) we develop a novel attention
mechanism to provide interpretable explanation for our model's decisions, (3)
we propose a novel sampling strategy exploring active learning to reduce the
burden of human annotation. OpenTag does not use any dictionary or hand-crafted
features as in prior works. Extensive experiments in real-life datasets in
different domains show that OpenTag with our active learning strategy discovers
new attribute values from as few as 150 annotated samples (reduction in 3.3x
amount of annotation effort) with a high F-score of 83%, outperforming
state-of-the-art models.Comment: Proceedings of the 24th ACM SIGKDD International Conference on
Knowledge Discovery and Data Mining, London, UK, August 19-23, 201
Similarity networks for classification: a case study in the Horse Colic problem
This paper develops a two-layer neural network in which the neuron model computes a user-defined similarity function between inputs and weights. The neuron transfer function is formed by composition of an adapted logistic function with the mean of the partial input-weight similarities. The resulting neuron model is capable of dealing directly with variables of potentially different nature (continuous, fuzzy, ordinal, categorical). There is also provision for missing values. The network is trained using a two-stage procedure very similar to that used to train a radial basis function (RBF) neural network. The network is compared to two types of RBF networks in a non-trivial dataset: the Horse Colic problem, taken as a case study and analyzed in detail.Postprint (published version
- …