236 research outputs found
A computationally and cognitively plausible model of supervised and unsupervised learning
Author version made available in accordance with the publisher's policy. "The final publication is available at link.springer.com”The issue of chance correction has been discussed for many decades in the context of
statistics, psychology and machine learning, with multiple measures being shown to
have desirable properties, including various definitions of Kappa or Correlation, and
the psychologically validated ΔP measures. In this paper, we discuss the relationships
between these measures, showing that they form part of a single family of measures,
and that using an appropriate measure can positively impact learning
Approximation and Relaxation Approaches for Parallel and Distributed Machine Learning
Large scale machine learning requires tradeoffs. Commonly this tradeoff has led practitioners to choose simpler, less powerful models, e.g. linear models, in order to process more training examples in a limited time. In this work, we introduce parallelism to the training of non-linear models by leveraging a different tradeoff--approximation. We demonstrate various techniques by which non-linear models can be made amenable to larger data sets and significantly more training parallelism by strategically introducing approximation in certain optimization steps.
For gradient boosted regression tree ensembles, we replace precise selection of tree splits with a coarse-grained, approximate split selection, yielding both faster sequential training and a significant increase in parallelism, in the distributed setting in particular. For metric learning with nearest neighbor classification, rather than explicitly train a neighborhood structure we leverage the implicit neighborhood structure induced by task-specific random forest classifiers, yielding a highly parallel method for metric learning. For support vector machines, we follow existing work to learn a reduced basis set with extremely high parallelism, particularly on GPUs, via existing linear algebra libraries.
We believe these optimization tradeoffs are widely applicable wherever machine learning is put in practice in large scale settings. By carefully introducing approximation, we also introduce significantly higher parallelism and consequently can process more training examples for more iterations than competing exact methods. While seemingly learning the model with less precision, this tradeoff often yields noticeably higher accuracy under a restricted training time budget
Active Multi-Field Learning for Spam Filtering
Ubiquitous spam messages cause a serious waste of time and resources. This paper addresses the practical spam filtering problem, and proposes a universal approach to fight with various spam messages. The proposed active multi-field learning approach is based on: 1) It is cost-sensitive to obtain a label for a real-world spam filter, which suggests an active learning idea; and 2) Different messages often have a similar multi-field text structure, which suggests a multi-field learning idea. The multi-field learning framework combines multiple results predicted from field classifiers by a novel compound weight, and each field classifier calculates the arithmetical average of multiple conditional probabilities predicted from feature strings according to a data structure of string-frequency index. Comparing the current variance of field classifying results with the historical variance, the active learner evaluates the classifying confidence and regards the more uncertain message as the more informative sample for which to request a label. The experimental results show that the proposed approach can achieve the state-of-the-art performance at greatly reduced label requirements both in email spam filtering and short text spam filtering. Our active multi-field learning performance, the standard (1-ROCA) % measurement, even exceeds the full feedback performance of some advanced individual classifying algorithm
Domain Adaptation Extreme Learning Machines for Drift Compensation in E-nose Systems
This paper addresses an important issue, known as sensor drift that behaves a
nonlinear dynamic property in electronic nose (E-nose), from the viewpoint of
machine learning. Traditional methods for drift compensation are laborious and
costly due to the frequent acquisition and labeling process for gases samples
recalibration. Extreme learning machines (ELMs) have been confirmed to be
efficient and effective learning techniques for pattern recognition and
regression. However, ELMs primarily focus on the supervised, semi-supervised
and unsupervised learning problems in single domain (i.e. source domain). To
our best knowledge, ELM with cross-domain learning capability has never been
studied. This paper proposes a unified framework, referred to as Domain
Adaptation Extreme Learning Machine (DAELM), which learns a robust classifier
by leveraging a limited number of labeled data from target domain for drift
compensation as well as gases recognition in E-nose systems, without loss of
the computational efficiency and learning ability of traditional ELM. In the
unified framework, two algorithms called DAELM-S and DAELM-T are proposed for
the purpose of this paper, respectively. In order to percept the differences
among ELM, DAELM-S and DAELM-T, two remarks are provided. Experiments on the
popular sensor drift data with multiple batches collected by E-nose system
clearly demonstrate that the proposed DAELM significantly outperforms existing
drift compensation methods without cumbersome measures, and also bring new
perspectives for ELM.Comment: 11 pages, 9 figures, to appear in IEEE Transactions on
Instrumentation and Measuremen
An overview of the main machine learning models - from theory to algorithms
Dissertation presented as the partial requirement for obtaining a Master's degree in Data Science and Advanced AnalyticsIn the context of solving highly complex problems, Artificial Intelligence shows an exponential
growth over the past years allowing the Machine Learning to augment and sometimes to outperform
the human learning. From driverless cars to automatic recommendation on Netflix, we are
surrounded by AI, even if we do not notice it. Furthermore, companies have recently adopted new
frameworks in their routines which are mainly composed by algorithms able to solve complex
problems in a short period of time.
The growth of AI technologies has been absolutely stunning and yes, it is only possible because
a sub-field of AI called Machine Learning is growing even faster. In a small scale, Machine Learning
may be seen as a simple system able to find patterns on data and learn from it. However, it is
precisely that learning process that in a large scale will allow machines to mimic the human behavior
and perform tasks that would eventually require human intelligence. Just for us to have an idea,
according to Forbes the global Machine Learning market was evaluated in 21B in 2024. Naturally, Machine Learning has become an attractive and
profitable scientific area that demands continuous learning since there is always something new
being discovered.
During the last decades, a huge number of algorithms have been proposed by the research
community, which sometimes may cause some confusion of how and when to use each one of them.
That is exactly what is pretended in this thesis, over the next chapters we are going to review the
main Machine Learning models and their respective advantages/disadvantages
- …