1,693 research outputs found
Summarizing Dialogic Arguments from Social Media
Online argumentative dialog is a rich source of information on popular
beliefs and opinions that could be useful to companies as well as governmental
or public policy agencies. Compact, easy to read, summaries of these dialogues
would thus be highly valuable. A priori, it is not even clear what form such a
summary should take. Previous work on summarization has primarily focused on
summarizing written texts, where the notion of an abstract of the text is well
defined. We collect gold standard training data consisting of five human
summaries for each of 161 dialogues on the topics of Gay Marriage, Gun Control
and Abortion. We present several different computational models aimed at
identifying segments of the dialogues whose content should be used for the
summary, using linguistic features and Word2vec features with both SVMs and
Bidirectional LSTMs. We show that we can identify the most important arguments
by using the dialog context with a best F-measure of 0.74 for gun control, 0.71
for gay marriage, and 0.67 for abortion.Comment: Proceedings of the 21th Workshop on the Semantics and Pragmatics of
Dialogue (SemDial 2017
3D Shape Knowledge Graph for Cross-domain and Cross-modal 3D Shape Retrieval
With the development of 3D modeling and fabrication, 3D shape retrieval has
become a hot topic. In recent years, several strategies have been put forth to
address this retrieval issue. However, it is difficult for them to handle
cross-modal 3D shape retrieval because of the natural differences between
modalities. In this paper, we propose an innovative concept, namely, geometric
words, which is regarded as the basic element to represent any 3D or 2D entity
by combination, and assisted by which, we can simultaneously handle
cross-domain or cross-modal retrieval problems. First, to construct the
knowledge graph, we utilize the geometric word as the node, and then use the
category of the 3D shape as well as the attribute of the geometry to bridge the
nodes. Second, based on the knowledge graph, we provide a unique way for
learning each entity's embedding. Finally, we propose an effective similarity
measure to handle the cross-domain and cross-modal 3D shape retrieval.
Specifically, every 3D or 2D entity could locate its geometric terms in the 3D
knowledge graph, which serve as a link between cross-domain and cross-modal
data. Thus, our approach can achieve the cross-domain and cross-modal 3D shape
retrieval at the same time. We evaluated our proposed method on the ModelNet40
dataset and ShapeNetCore55 dataset for both the 3D shape retrieval task and
cross-domain 3D shape retrieval task. The classic cross-modal dataset (MI3DOR)
is utilized to evaluate cross-modal 3D shape retrieval. Experimental results
and comparisons with state-of-the-art methods illustrate the superiority of our
approach
- …