94,178 research outputs found
Regression on feature projections
Cataloged from PDF version of article.This paper describes a machine learning method, called Regression on Feature Projections (RFP), for predicting a real-valued target
feature, given the values of multiple predictive features. In RFP training is based on simply storing the projections of the training instances on
each feature separately. Prediction of the target value for a query point is obtained through two averaging procedures executed sequentially.
The ®rst averaging process is to ®nd the individual predictions of features by using the K-Nearest Neighbor (KNN) algorithm. The second
averaging process combines the predictions of all features. During the ®rst averaging step, each feature is associated with a weight in order to
determine the prediction ability of the feature at the local query point. The weights, found for each local query point, are used in the second
prediction step and enforce the method to have an adaptive or context-sensitive nature. We have compared RFP with KNN and the rule
based-regression algorithms. Results on real data sets show that RFP achieves better or comparable accuracy and is faster than both KNN and
Rule-based regression algorithms. (C)2000 Elsevier Science B.V. All rights reserved
Application of Memory-Based Collaborative Filtering to Predict Fantasy Points of NFL Quarterbacks
Subjective expert projections have been traditionally used to predict points in fantasy football, while machine prediction applications are limited. Memory-based collaborative filtering has been widely used in recommender system domain to predict ratings and recommend items. In this study, user-based and item-based collaborative filtering were explored and implemented to predict the weekly statistics and fantasy points of NFL quarterbacks. The predictions from three seasons were compared against expert projections. On both weekly statistics and total fantasy points, the implementations could not make significantly better predictions than experts.However, the prediction from the implementation improved the accuracy of other regression models when used as additional feature
Bellman Error Based Feature Generation using Random Projections on Sparse Spaces
We address the problem of automatic generation of features for value function
approximation. Bellman Error Basis Functions (BEBFs) have been shown to improve
the error of policy evaluation with function approximation, with a convergence
rate similar to that of value iteration. We propose a simple, fast and robust
algorithm based on random projections to generate BEBFs for sparse feature
spaces. We provide a finite sample analysis of the proposed method, and prove
that projections logarithmic in the dimension of the original space are enough
to guarantee contraction in the error. Empirical results demonstrate the
strength of this method
A Comparison of Machine Learning Methods for Cross-Domain Few-Shot Learning
We present an empirical evaluation of machine learning algorithms in cross-domain few-shot learning based on a fixed pre-trained feature extractor. Experiments were performed in five target domains (CropDisease, EuroSAT, Food101, ISIC and ChestX) and using two feature extractors: a ResNet10 model trained on a subset of ImageNet known as miniImageNet and a ResNet152 model trained on the ILSVRC 2012 subset of ImageNet. Commonly used machine learning algorithms including logistic regression, support vector machines, random forests, nearest neighbour classification, naïve Bayes, and linear and quadratic discriminant analysis were evaluated on the extracted feature vectors. We also evaluated classification accuracy when subjecting the feature vectors to normalisation using p-norms. Algorithms originally developed for the classification of gene expression data—the nearest shrunken centroid algorithm and LDA ensembles obtained with random projections—were also included in the experiments, in addition to a cosine similarity classifier that has recently proved popular in few-shot learning. The results enable us to identify algorithms, normalisation methods and pre-trained feature extractors that perform well in cross-domain few-shot learning. We show that the cosine similarity classifier and ℓ² -regularised 1-vs-rest logistic regression are generally the best-performing algorithms. We also show that algorithms such as LDA yield consistently higher accuracy when applied to ℓ² -normalised feature vectors. In addition, all classifiers generally perform better when extracting feature vectors using the ResNet152 model instead of the ResNet10 model
- …