8,025 research outputs found
QDEE: Question Difficulty and Expertise Estimation in Community Question Answering Sites
In this paper, we present a framework for Question Difficulty and Expertise
Estimation (QDEE) in Community Question Answering sites (CQAs) such as Yahoo!
Answers and Stack Overflow, which tackles a fundamental challenge in
crowdsourcing: how to appropriately route and assign questions to users with
the suitable expertise. This problem domain has been the subject of much
research and includes both language-agnostic as well as language conscious
solutions. We bring to bear a key language-agnostic insight: that users gain
expertise and therefore tend to ask as well as answer more difficult questions
over time. We use this insight within the popular competition (directed) graph
model to estimate question difficulty and user expertise by identifying key
hierarchical structure within said model. An important and novel contribution
here is the application of "social agony" to this problem domain. Difficulty
levels of newly posted questions (the cold-start problem) are estimated by
using our QDEE framework and additional textual features. We also propose a
model to route newly posted questions to appropriate users based on the
difficulty level of the question and the expertise of the user. Extensive
experiments on real world CQAs such as Yahoo! Answers and Stack Overflow data
demonstrate the improved efficacy of our approach over contemporary
state-of-the-art models. The QDEE framework also allows us to characterize user
expertise in novel ways by identifying interesting patterns and roles played by
different users in such CQAs.Comment: Accepted in the Proceedings of the 12th International AAAI Conference
on Web and Social Media (ICWSM 2018). June 2018. Stanford, CA, US
Simplifying Sparse Expert Recommendation by Revisiting Graph Diffusion
Community Question Answering (CQA) websites have become valuable knowledge
repositories where individuals exchange information by asking and answering
questions. With an ever-increasing number of questions and high migration of
users in and out of communities, a key challenge is to design effective
strategies for recommending experts for new questions. In this paper, we
propose a simple graph-diffusion expert recommendation model for CQA, that can
outperform state-of-the art deep learning representatives and collaborative
models. Our proposed method learns users' expertise in the context of both
semantic and temporal information to capture their changing interest and
activity levels with time. Experiments on five real-world datasets from the
Stack Exchange network demonstrate that our approach outperforms competitive
baseline methods. Further, experiments on cold-start users (users with a
limited historical record) show our model achieves an average of ~ 30%
performance gain compared to the best baseline method
Four Degrees of Separation, Really
We recently measured the average distance of users in the Facebook graph,
spurring comments in the scientific community as well as in the general press
("Four Degrees of Separation"). A number of interesting criticisms have been
made about the meaningfulness, methods and consequences of the experiment we
performed. In this paper we want to discuss some methodological aspects that we
deem important to underline in the form of answers to the questions we have
read in newspapers, magazines, blogs, or heard from colleagues. We indulge in
some reflections on the actual meaning of "average distance" and make a number
of side observations showing that, yes, 3.74 "degrees of separation" are really
few
- …