3 research outputs found
Quantifying Aspect Bias in Ordinal Ratings using a Bayesian Approach
User opinions expressed in the form of ratings can influence an individual's
view of an item. However, the true quality of an item is often obfuscated by
user biases, and it is not obvious from the observed ratings the importance
different users place on different aspects of an item. We propose a
probabilistic modeling of the observed aspect ratings to infer (i) each user's
aspect bias and (ii) latent intrinsic quality of an item. We model multi-aspect
ratings as ordered discrete data and encode the dependency between different
aspects by using a latent Gaussian structure. We handle the
Gaussian-Categorical non-conjugacy using a stick-breaking formulation coupled
with P\'{o}lya-Gamma auxiliary variable augmentation for a simple, fully
Bayesian inference. On two real world datasets, we demonstrate the predictive
ability of our model and its effectiveness in learning explainable user biases
to provide insights towards a more reliable product quality estimation.Comment: Accepted for publication in IJCAI 201
The Latent Topic Block Model for the Co-Clustering of Textual Interaction Data
International audienceIn this paper, we consider textual interaction data involving two disjoint sets of individuals/objects. An example of such data is given by the reviews on web platforms (e.g. Amazon, TripAdvisor, etc.) where buyers comment on products/services they bought. We develop a new generative model, the latent topic block model (LTBM), along with an inference algorithm to simultaneously partition the elements of each set, accounting for the textual information. The estimation of the model parameters is performed via a variational version of the expectation maximization (EM) algorithm. A model selection criterion is formally obtained to estimate the number of partitions. Numerical experiments on simulated data are carried out to highlight the main features of the estimation procedure. Two real-world datasets are finally employed to show the usefulness of the proposed approach
Latent Dirichlet Bayesian co-clustering
Abstract. Co-clustering has emerged as an important technique for mining contingency data matrices. However, almost all existing coclustering algorithms are hard partitioning, assigning each row and column of the data matrix to one cluster. Recently a Bayesian co-clustering approach has been proposed which allows a probability distribution membership in row and column clusters. The approach uses variational inference for parameter estimation. In this work, we modify the Bayesian co-clustering model, and use collapsed Gibbs sampling and collapsed variational inference for parameter estimation. Our empirical evaluation on real data sets shows that both collapsed Gibbs sampling and collapsed variational inference are able to find more accurate likelihood estimates than the standard variational Bayesian co-clustering approach