520,837 research outputs found
GAMLSS for high-dimensional data – a flexible approach based on boosting
Generalized additive models for location, scale and shape (GAMLSS) are a popular semi-parametric modelling approach that, in contrast to conventional GAMs, regress not only the expected mean but every distribution parameter (e.g. location, scale and shape) to a set of covariates. Current fitting procedures for GAMLSS are infeasible for high-dimensional data setups and require variable selection based on (potentially problematic) information criteria. The present work describes a boosting algorithm for high-dimensional GAMLSS that was developed to overcome these limitations. Specifically, the new algorithm was designed to allow the simultaneous estimation of predictor effects and variable selection. The proposed algorithm was applied to data of the Munich Rental Guide, which is used by
landlords and tenants as a reference for the average rent of a flat depending on its characteristics and spatial features. The net-rent predictions that resulted from the high-dimensional GAMLSS were found to be highly competitive while covariate-specific prediction intervals showed a major improvement over classical GAMs
A Feature Selection Method for Multivariate Performance Measures
Feature selection with specific multivariate performance measures is the key
to the success of many applications, such as image retrieval and text
classification. The existing feature selection methods are usually designed for
classification error. In this paper, we propose a generalized sparse
regularizer. Based on the proposed regularizer, we present a unified feature
selection framework for general loss functions. In particular, we study the
novel feature selection paradigm by optimizing multivariate performance
measures. The resultant formulation is a challenging problem for
high-dimensional data. Hence, a two-layer cutting plane algorithm is proposed
to solve this problem, and the convergence is presented. In addition, we adapt
the proposed method to optimize multivariate measures for multiple instance
learning problems. The analyses by comparing with the state-of-the-art feature
selection methods show that the proposed method is superior to others.
Extensive experiments on large-scale and high-dimensional real world datasets
show that the proposed method outperforms -SVM and SVM-RFE when choosing a
small subset of features, and achieves significantly improved performances over
SVM in terms of -score
A Taguchi method application for the part routing selection in Generalized Group Technology: A case Study
Cellular manufacturing (CM) is an important application of group technology (GT) that can be used to enhance both flexibility and efficiency in today’s small-to-medium lot production environment. The crucial step in the design of a CM system is the cell formation (CF) problem which involves grouping parts into families and machines into cells. The CF problem are increasingly complicated if parts are assigned with alternative routings (known as generalized Group Technology problem). In most of the previous works, the route selection problem and CF problem were formulated in a single model which is not practical for solving large-scale problems. We suggest that better solution could be obtained by formulating and solving them separately in two different problems. The aim of this case study is to apply Taguchi method for the route selection problem as an optimization technique to get back to the simple CF problem which can be solved by any of the numerous CF procedures. In addition the main effect of each part and analysis of variance (ANOVA) are introduced as a sensitivity analysis aspect that is completely ignored in previous research.Cellular Manufacturing; generalized Group Technology; route selection problem; Taguchi method; ANOVA; sensitivity analysis
Rational Krylov approximation of matrix functions: Numerical methods and optimal pole selection
Matrix functions are a central topic of linear algebra, and problems of their numerical approximation appear increasingly often in scientific computing. We review various rational Krylov methods for the computation of large-scale matrix functions. Emphasis is put on the rational Arnoldi method and variants thereof, namely, the extended Krylov subspace method and the shift-and-invert Arnoldi method, but we also discuss the nonorthogonal generalized Leja point (or PAIN) method. The issue of optimal pole selection for rational Krylov methods applied for approximating the resolvent and exponential function, and functions of Markov type, is treated in some detail
- …