405 research outputs found
Men Are from Mars, Women Are from Venus: Evaluation and Modelling of Verbal Associations
We present a quantitative analysis of human word association pairs and study
the types of relations presented in the associations. We put our main focus on
the correlation between response types and respondent characteristics such as
occupation and gender by contrasting syntagmatic and paradigmatic associations.
Finally, we propose a personalised distributed word association model and show
the importance of incorporating demographic factors into the models commonly
used in natural language processing.Comment: AIST 2017 camera-read
A Latent Dirichlet Framework for Relevance Modeling
Abstract. Relevance-based language models operate by estimating the probabilities of observing words in documents relevant (or pseudo relevant) to a topic. However, these models assume that if a document is relevant to a topic, then all tokens in the document are relevant to that topic. This could limit model robustness and effectiveness. In this study, we propose a Latent Dirichlet relevance model, which relaxes this assumption. Our approach derives from current research on Latent Dirichlet Allocation (LDA) topic models. LDA has been extensively explored, especially for generating a set of topics from a corpus. A key attraction is that in LDA a document may be about several topics. LDA itself, however, has a limitation that is also addressed in our work. Topics generated by LDA from a corpus are synthetic, i.e., they do not necessarily correspond to topics identified by humans for the same corpus. In contrast, our model explicitly considers the relevance relationships between documents and given topics (queries). Thus unlike standard LDA, our model is directly applicable to goals such as relevance feedback for query modification and text classification, where topics (classes and queries) are provided upfront. Thus although the focus of our paper is on improving relevance-based language models, in effect our approach bridges relevance-based language models and LDA addressing limitations of both. Finally, we propose an idea that takes advantage of “bagof-words” assumption to reduce the complexity of Gibbs sampling based learning algorithm
Using Conversation Topics for Predicting Therapy Outcomes in Schizophrenia
Previous research shows that aspects of doctor-patient communication in therapy can predict patient symptoms, satisfaction and future adherence to treatment (a significant problem with conditions such as schizophrenia). However, automatic prediction has so far shown success only when based on low-level lexical features, and it is unclear how well these can generalize to new data, or whether their effectiveness is due to their capturing aspects of style, structure or content. Here, we examine the use of topic as a higher-level measure of content, more likely to generalize and to have more explanatory power. Investigations show that while topics predict some important factors such as patient satisfaction and ratings of therapy quality, they lack the full predictive power of lower-level features. For some factors, unsupervised methods produce models comparable to manual annotation
Enhancing complex-network synchronization
Heterogeneity in the degree (connectivity) distribution has been shown to
suppress synchronization in networks of symmetrically coupled oscillators with
uniform coupling strength (unweighted coupling). Here we uncover a condition
for enhanced synchronization in directed networks with weighted coupling. We
show that, in the optimum regime, synchronizability is solely determined by the
average degree and does not depend on the system size and the details of the
degree distribution. In scale-free networks, where the average degree may
increase with heterogeneity, synchronizability is drastically enhanced and may
become positively correlated with heterogeneity, while the overall cost
involved in the network coupling is significantly reduced as compared to the
case of unweighted coupling.Comment: 4 pages, 3 figure
Modeling individual differences with Dirichlet processes
We introduce a Bayesian framework for modeling individual differences, in which subjects are assumed to belong to one of a potentially infinite number of groups. In this model, the groups observed in any particular data set are not viewed as a fixed set that fully explain the variation between individuals, but rather as representatives of a latent, arbitrarily rich structure. As more people are seen, the number of observed groups is allowed to grow, as more details about the individual differences are revealed. We use the Dirichlet process – a distribution widely used in nonparametric Bayesian statistics – to define a prior for the model, allowing us to learn flexible parameter distributions without overfitting the data, or requiring the complex computations typically required for determining the dimensionality of a model. As an initial demonstration of the approach, we present an application of the method to categorization data.Daniel J. Navarro, Thomas L. Griffiths, Mark Steyvers, Michael D. Le
- …