27 research outputs found

    Regularizing soft decision trees

    Get PDF
    Recently, we have proposed a new decision tree family called soft decision trees where a node chooses both its left and right children with different probabilities as given by a gating function, different from a hard decision node which chooses one of the two. In this paper, we extend the original algorithm by introducing local dimension reduction via L-1 and L-2 regularization for feature selection and smoother fitting. We compare our novel approach with the standard decision tree algorithms over 27 classification data sets. We see that both regularized versions have similar generalization ability with less complexity in terms of number of nodes, where L-2 seems to work slightly better than L-1.Publisher's VersionAuthor Post Prin

    A Comparison of Model Aggregation Methods for Regression

    No full text
    Combining machine learning models is a means of improving overall accuracy. Various algorithms have been proposed to create aggregate models from other models, and two popular examples for classification are Bagging and AdaBoost

    Gabor Wavelet Based Pose Estimation For Face

    No full text
    One of the major difficulties in face recognition systems is the in-depth pose variation problem. Most face recognition approaches assume that the pose of the face is known. In this work, we use a variation of Gabor wavelet transform for the representation of human face images to efficiently solve the pose estimation problem. Parameters of the Gabor wavelets, namely frequency and orientation, are adjusted to gain better performance. Principal Component Analysis is performed to reduce the dimensionality without a significant loss in the performance. Our results show that Gabor wavelet based filtering of images improves the performance of the pose estimation module

    Statistical Tests Using Hinge/ε-Sensitive Loss

    No full text

    A model for predicting drying time period of wool yarn bobbins using computational intelligence techniques

    No full text
    In this study, a predictive model has been developed using computational intelligence techniques for the prediction of drying time in the wool yarn bobbin drying process. The bobbin drying process is influenced by various drying parameters, 19 of which were used as input variables in the dataset. These parameters affect the drying time of yarn bobbins, which is considered as the target variable. The dataset, which consists of these input and target variables, was collected from an experimental yarn bobbin drying system. Firstly, the most effective input variables on the target variable, named as the best feature subset of the dataset, were investigated by using a filter-based feature selection method. As a result, the most important five parameters were obtained as the best feature subset. Afterwards, the most successful method that can predict the drying time of wool yarn bobbins with the highest accuracy was explored amongst the 16 computational intelligence methods for the best feature subset. Finally, the best performance has been found by the REP tree method, which achieved minimum error and time taken to build the model.TUBITAKTurkiye Bilimsel ve Teknolojik Arastirma Kurumu (TUBITAK) [108M274]This work was supported by TUBITAK (grant number 108M274)

    Optimal Gabor Kernel Location Selection For Face Recognition

    No full text
    In local feature--based face recognition systems, the topographical locations of feature extractors directly affect the discriminative power of a recognizer. Better recognition accuracy can be achieved by the determination of the positions of salient image locations. Most of the facial feature selection algorithms in the literature work with two assumptions: one, that the importance of each feature is independent of the other features, and two, that the kernels should be located at fiducial points. Under these assumption, one can only get a sub--optimal solution. In this paper, we present a methodology that tries to overcome this problem by relaxing the two assumptions using a formalism of subset selection problem. We use a number of feature selection algorithms and a genetic algorithm. Comparative results on the FERET dataset confirm the viability of our approach

    Mixtures of Large Margin Nearest Neighbor Classifiers

    No full text
    Abstract. The accuracy of the k-nearest neighbor algorithm depends on the distance function used to measure similarity between instances. Methods have been proposed in the literature to learn a good distance function from a labelled training set. One such method is the large margin nearest neighbor classifier that learns a global Mahalanobis distance. We propose a mixture of such classifiers where a gating function divides the input space into regions and a separate distance function is learned in each region in a lower dimensional manifold. We show that such an extension improves accuracy and allows visualization
    corecore