165 research outputs found
Hierarchical Object Parsing from Structured Noisy Point Clouds
Object parsing and segmentation from point clouds are challenging tasks
because the relevant data is available only as thin structures along object
boundaries or other features, and is corrupted by large amounts of noise. To
handle this kind of data, flexible shape models are desired that can accurately
follow the object boundaries. Popular models such as Active Shape and Active
Appearance models lack the necessary flexibility for this task, while recent
approaches such as the Recursive Compositional Models make model
simplifications in order to obtain computational guarantees. This paper
investigates a hierarchical Bayesian model of shape and appearance in a
generative setting. The input data is explained by an object parsing layer,
which is a deformation of a hidden PCA shape model with Gaussian prior. The
paper also introduces a novel efficient inference algorithm that uses informed
data-driven proposals to initialize local searches for the hidden variables.
Applied to the problem of object parsing from structured point clouds such as
edge detection images, the proposed approach obtains state of the art parsing
errors on two standard datasets without using any intensity information.Comment: 13 pages, 16 figure
Are screening methods useful in feature selection? An empirical study
Filter or screening methods are often used as a preprocessing step for
reducing the number of variables used by a learning algorithm in obtaining a
classification or regression model. While there are many such filter methods,
there is a need for an objective evaluation of these methods. Such an
evaluation is needed to compare them with each other and also to answer whether
they are at all useful, or a learning algorithm could do a better job without
them. For this purpose, many popular screening methods are partnered in this
paper with three regression learners and five classification learners and
evaluated on ten real datasets to obtain accuracy criteria such as R-square and
area under the ROC curve (AUC). The obtained results are compared through curve
plots and comparison tables in order to find out whether screening methods help
improve the performance of learning algorithms and how they fare with each
other. Our findings revealed that the screening methods were useful in
improving the prediction of the best learner on two regression and two
classification datasets out of the ten datasets evaluated.Comment: 29 pages, 4 figures, 21 table
Generating Compact Tree Ensembles via Annealing
Tree ensembles are flexible predictive models that can capture relevant
variables and to some extent their interactions in a compact and interpretable
manner. Most algorithms for obtaining tree ensembles are based on versions of
boosting or Random Forest. Previous work showed that boosting algorithms
exhibit a cyclic behavior of selecting the same tree again and again due to the
way the loss is optimized. At the same time, Random Forest is not based on loss
optimization and obtains a more complex and less interpretable model. In this
paper we present a novel method for obtaining compact tree ensembles by growing
a large pool of trees in parallel with many independent boosting threads and
then selecting a small subset and updating their leaf weights by loss
optimization. We allow for the trees in the initial pool to have different
depths which further helps with generalization. Experiments on real datasets
show that the obtained model has usually a smaller loss than boosting, which is
also reflected in a lower misclassification error on the test set.Comment: Comparison with Random Forest included in the results sectio
Learning Mixtures of Bernoulli Templates by Two-Round EM with Performance Guarantee
Dasgupta and Shulman showed that a two-round variant of the EM algorithm can
learn mixture of Gaussian distributions with near optimal precision with high
probability if the Gaussian distributions are well separated and if the
dimension is sufficiently high. In this paper, we generalize their theory to
learning mixture of high-dimensional Bernoulli templates. Each template is a
binary vector, and a template generates examples by randomly switching its
binary components independently with a certain probability. In computer vision
applications, a binary vector is a feature map of an image, where each binary
component indicates whether a local feature or structure is present or absent
within a certain cell of the image domain. A Bernoulli template can be
considered as a statistical model for images of objects (or parts of objects)
from the same category. We show that the two-round EM algorithm can learn
mixture of Bernoulli templates with near optimal precision with high
probability, if the Bernoulli templates are sufficiently different and if the
number of features is sufficiently high. We illustrate the theoretical results
by synthetic and real examples.Comment: 27 pages, 8 figure
Feature Selection with Annealing for Forecasting Financial Time Series
Stock market and cryptocurrency forecasting is very important to investors as
they aspire to achieve even the slightest improvement to their buy or hold
strategies so that they may increase profitability. However, obtaining accurate
and reliable predictions is challenging, noting that accuracy does not equate
to reliability, especially when financial time-series forecasting is applied
owing to its complex and chaotic tendencies. To mitigate this complexity, this
study provides a comprehensive method for forecasting financial time series
based on tactical input output feature mapping techniques using machine
learning (ML) models. During the prediction process, selecting the relevant
indicators is vital to obtaining the desired results. In the financial field,
limited attention has been paid to this problem with ML solutions. We
investigate the use of feature selection with annealing (FSA) for the first
time in this field, and we apply the least absolute shrinkage and selection
operator (Lasso) method to select the features from more than 1,000 candidates
obtained from 26 technical classifiers with different periods and lags. Boruta
(BOR) feature selection, a wrapper method, is used as a baseline for
comparison. Logistic regression (LR), extreme gradient boosting (XGBoost), and
long short-term memory (LSTM) are then applied to the selected features for
forecasting purposes using 10 different financial datasets containing
cryptocurrencies and stocks. The dependent variables consisted of daily
logarithmic returns and trends. The mean-squared error for regression, area
under the receiver operating characteristic curve, and classification accuracy
were used to evaluate model performance, and the statistical significance of
the forecasting results was tested using paired t-tests. Experiments indicate
that the FSA algorithm increased the performance of ML models, regardless of
problem type.Comment: 37 pages, 1 figures and 12 table
- …