520,837 research outputs found

    GAMLSS for high-dimensional data – a flexible approach based on boosting

    Get PDF
    Generalized additive models for location, scale and shape (GAMLSS) are a popular semi-parametric modelling approach that, in contrast to conventional GAMs, regress not only the expected mean but every distribution parameter (e.g. location, scale and shape) to a set of covariates. Current fitting procedures for GAMLSS are infeasible for high-dimensional data setups and require variable selection based on (potentially problematic) information criteria. The present work describes a boosting algorithm for high-dimensional GAMLSS that was developed to overcome these limitations. Specifically, the new algorithm was designed to allow the simultaneous estimation of predictor effects and variable selection. The proposed algorithm was applied to data of the Munich Rental Guide, which is used by landlords and tenants as a reference for the average rent of a flat depending on its characteristics and spatial features. The net-rent predictions that resulted from the high-dimensional GAMLSS were found to be highly competitive while covariate-specific prediction intervals showed a major improvement over classical GAMs

    A Feature Selection Method for Multivariate Performance Measures

    Full text link
    Feature selection with specific multivariate performance measures is the key to the success of many applications, such as image retrieval and text classification. The existing feature selection methods are usually designed for classification error. In this paper, we propose a generalized sparse regularizer. Based on the proposed regularizer, we present a unified feature selection framework for general loss functions. In particular, we study the novel feature selection paradigm by optimizing multivariate performance measures. The resultant formulation is a challenging problem for high-dimensional data. Hence, a two-layer cutting plane algorithm is proposed to solve this problem, and the convergence is presented. In addition, we adapt the proposed method to optimize multivariate measures for multiple instance learning problems. The analyses by comparing with the state-of-the-art feature selection methods show that the proposed method is superior to others. Extensive experiments on large-scale and high-dimensional real world datasets show that the proposed method outperforms l1l_1-SVM and SVM-RFE when choosing a small subset of features, and achieves significantly improved performances over SVMperf^{perf} in terms of F1F_1-score

    A Taguchi method application for the part routing selection in Generalized Group Technology: A case Study

    Get PDF
    Cellular manufacturing (CM) is an important application of group technology (GT) that can be used to enhance both flexibility and efficiency in today’s small-to-medium lot production environment. The crucial step in the design of a CM system is the cell formation (CF) problem which involves grouping parts into families and machines into cells. The CF problem are increasingly complicated if parts are assigned with alternative routings (known as generalized Group Technology problem). In most of the previous works, the route selection problem and CF problem were formulated in a single model which is not practical for solving large-scale problems. We suggest that better solution could be obtained by formulating and solving them separately in two different problems. The aim of this case study is to apply Taguchi method for the route selection problem as an optimization technique to get back to the simple CF problem which can be solved by any of the numerous CF procedures. In addition the main effect of each part and analysis of variance (ANOVA) are introduced as a sensitivity analysis aspect that is completely ignored in previous research.Cellular Manufacturing; generalized Group Technology; route selection problem; Taguchi method; ANOVA; sensitivity analysis

    Rational Krylov approximation of matrix functions: Numerical methods and optimal pole selection

    Get PDF
    Matrix functions are a central topic of linear algebra, and problems of their numerical approximation appear increasingly often in scientific computing. We review various rational Krylov methods for the computation of large-scale matrix functions. Emphasis is put on the rational Arnoldi method and variants thereof, namely, the extended Krylov subspace method and the shift-and-invert Arnoldi method, but we also discuss the nonorthogonal generalized Leja point (or PAIN) method. The issue of optimal pole selection for rational Krylov methods applied for approximating the resolvent and exponential function, and functions of Markov type, is treated in some detail
    • …
    corecore