20 research outputs found
Linear Bandits with Feature Feedback
This paper explores a new form of the linear bandit problem in which the
algorithm receives the usual stochastic rewards as well as stochastic feedback
about which features are relevant to the rewards, the latter feedback being the
novel aspect. The focus of this paper is the development of new theory and
algorithms for linear bandits with feature feedback. We show that linear
bandits with feature feedback can achieve regret over time horizon that
scales like , without prior knowledge of which features are relevant
nor the number of relevant features. In comparison, the regret of
traditional linear bandits is , where is the total number of
(relevant and irrelevant) features, so the improvement can be dramatic if . The computational complexity of the new algorithm is proportional to
rather than , making it much more suitable for real-world applications
compared to traditional linear bandits. We demonstrate the performance of the
new algorithm with synthetic and real human-labeled data
Discovering fine-grained sentiment with latent variable structured prediction models
In this paper we investigate the use of latent variable structured prediction models for fine-grained sentiment analysis in the common situation where only coarse-grained supervision is available. Specifically, we show how sentence-level sentiment labels can be effectively learned from document-level supervision using hidden conditional random fields (HCRFs). Experiments show that this technique reduces sentence classification errors by 22\% relative to using a lexicon and by 13\% relative to machine-learning baselines.
We provide a comprehensible description of the proposed probabilistic model and the features employed. Further, we describe the construction of a manually annotated test set, which was used in a thorough empirical investigation of the performance of the proposed model
Don't Let Me Be Misunderstood: Comparing Intentions and Perceptions in Online Discussions
Discourse involves two perspectives: a person's intention in making an
utterance and others' perception of that utterance. The misalignment between
these perspectives can lead to undesirable outcomes, such as misunderstandings,
low productivity and even overt strife. In this work, we present a
computational framework for exploring and comparing both perspectives in online
public discussions.
We combine logged data about public comments on Facebook with a survey of
over 16,000 people about their intentions in writing these comments or about
their perceptions of comments that others had written. Unlike previous studies
of online discussions that have largely relied on third-party labels to
quantify properties such as sentiment and subjectivity, our approach also
directly captures what the speakers actually intended when writing their
comments. In particular, our analysis focuses on judgments of whether a comment
is stating a fact or an opinion, since these concepts were shown to be often
confused.
We show that intentions and perceptions diverge in consequential ways. People
are more likely to perceive opinions than to intend them, and linguistic cues
that signal how an utterance is intended can differ from those that signal how
it will be perceived. Further, this misalignment between intentions and
perceptions can be linked to the future health of a conversation: when a
comment whose author intended to share a fact is misperceived as sharing an
opinion, the subsequent conversation is more likely to derail into uncivil
behavior than when the comment is perceived as intended. Altogether, these
findings may inform the design of discussion platforms that better promote
positive interactions.Comment: Proceedings of The Web Conference (WWW) 202