20,072 research outputs found

    A comparison of two-stage segmentation methods for choice-based conjoint data: a simulation study.

    Get PDF
    Due to the increasing interest in market segmentation in modern marketing research, several methods for dealing with consumer heterogeneity and for revealing market segments have been described in the literature. In this study, the authors compare eight two-stage segmentation methods that aim to uncover consumer segments by classifying subject-specific indicator values. Four different indicators are used as a segmentation basis. The forces, which are subject-aggregated gradient values of the likelihood function, and the dfbetas, an outlier detection measure, are two indicators that express a subject’s effect on the estimation of the aggregate partworths in the conditional logit model. Although the conditional logit model is generally estimated at the aggregate level, this research obtains individual-level partworth estimates for segmentation purposes. The respondents’ raw choices are the final indicator values. The authors classify the indicators by means of cluster analysis and latent class models. The goal of the study is to compare the segmentation performance of the methods with respect to their success rate, membership recovery and segment mean parameter recovery. With regard to the individual-level estimates, the authors obtain poor segmentation results both with cluster and latent class analysis. The cluster methods based on the forces, the dfbetas and the choices yield good and similar results. Classification of the forces and the dfbetas deteriorates with the use of latent class analysis, whereas latent class modeling of the choices outperforms its cluster counterpart.Two-stage segmentation methods; Choice-based conjoint analysis; Conditional logit model; Market segmentation; Latent class analysis;

    Producing power-law distributions and damping word frequencies with two-stage language models

    Get PDF
    Standard statistical models of language fail to capture one of the most striking properties of natural languages: the power-law distribution in the frequencies of word tokens. We present a framework for developing statisticalmodels that can generically produce power laws, breaking generativemodels into two stages. The first stage, the generator, can be any standard probabilistic model, while the second stage, the adaptor, transforms the word frequencies of this model to provide a closer match to natural language. We show that two commonly used Bayesian models, the Dirichlet-multinomial model and the Dirichlet process, can be viewed as special cases of our framework. We discuss two stochastic processes-the Chinese restaurant process and its two-parameter generalization based on the Pitman-Yor process-that can be used as adaptors in our framework to produce power-law distributions over word frequencies. We show that these adaptors justify common estimation procedures based on logarithmic or inverse-power transformations of empirical frequencies. In addition, taking the Pitman-Yor Chinese restaurant process as an adaptor justifies the appearance of type frequencies in formal analyses of natural language and improves the performance of a model for unsupervised learning of morphology.48 page(s

    Adaptive Multidimensional Scaling: The Spatial Representation of Brand Consideration and Dissimilarity Judgments

    Get PDF
    We propose Adaptive Multidimensional Scaling (AMDS) for simultaneously deriving a brand map and market segments using consumer data on cognitive decision sets and brand dissimilarities.In AMDS, the judgment task is adapted to the individual respondent: dissimilarity judgments are collected only for those brands within a consumers' awareness set.Thus, respondent fatigue and subjects' unfamiliarity with any subset of the brands are circumvented; thereby improving the validity of the dissimilarity data obtained, as well as the multidimensional spatial structure derived.Estimation of the AMDS model results in a spatial map in which the brands and derived segments of consumers are jointly represented as points.The closer a brand is positioned to a segment's ideal brand, the higher the probability that the brand is considered and chosen.An assumption underlying this model representation is that brands within a consumers' consideration set are relatively similar.In an experiment with 200 subjects and 4 product categories, this assumption is validated.We illustrate adaptive multidimensional scaling on commercial data for 20 midsize car brands evaluated by 212 members of a consumer panel.Potential applications of the method and future research opportunities are discussed.scaling;brands;market segmentation

    Event-Based Modeling with High-Dimensional Imaging Biomarkers for Estimating Spatial Progression of Dementia

    Full text link
    Event-based models (EBM) are a class of disease progression models that can be used to estimate temporal ordering of neuropathological changes from cross-sectional data. Current EBMs only handle scalar biomarkers, such as regional volumes, as inputs. However, regional aggregates are a crude summary of the underlying high-resolution images, potentially limiting the accuracy of EBM. Therefore, we propose a novel method that exploits high-dimensional voxel-wise imaging biomarkers: n-dimensional discriminative EBM (nDEBM). nDEBM is based on an insight that mixture modeling, which is a key element of conventional EBMs, can be replaced by a more scalable semi-supervised support vector machine (SVM) approach. This SVM is used to estimate the degree of abnormality of each region which is then used to obtain subject-specific disease progression patterns. These patterns are in turn used for estimating the mean ordering by fitting a generalized Mallows model. In order to validate the biomarker ordering obtained using nDEBM, we also present a framework for Simulation of Imaging Biomarkers' Temporal Evolution (SImBioTE) that mimics neurodegeneration in brain regions. SImBioTE trains variational auto-encoders (VAE) in different brain regions independently to simulate images at varying stages of disease progression. We also validate nDEBM clinically using data from the Alzheimer's Disease Neuroimaging Initiative (ADNI). In both experiments, nDEBM using high-dimensional features gave better performance than state-of-the-art EBM methods using regional volume biomarkers. This suggests that nDEBM is a promising approach for disease progression modeling.Comment: IPMI 201

    Multi-Source Neural Variational Inference

    Full text link
    Learning from multiple sources of information is an important problem in machine-learning research. The key challenges are learning representations and formulating inference methods that take into account the complementarity and redundancy of various information sources. In this paper we formulate a variational autoencoder based multi-source learning framework in which each encoder is conditioned on a different information source. This allows us to relate the sources via the shared latent variables by computing divergence measures between individual source's posterior approximations. We explore a variety of options to learn these encoders and to integrate the beliefs they compute into a consistent posterior approximation. We visualise learned beliefs on a toy dataset and evaluate our methods for learning shared representations and structured output prediction, showing trade-offs of learning separate encoders for each information source. Furthermore, we demonstrate how conflict detection and redundancy can increase robustness of inference in a multi-source setting.Comment: AAAI 2019, Association for the Advancement of Artificial Intelligence (AAAI) 201
    corecore