532 research outputs found
Fast and Accurate Neural Word Segmentation for Chinese
Neural models with minimal feature engineering have achieved competitive
performance against traditional methods for the task of Chinese word
segmentation. However, both training and working procedures of the current
neural models are computationally inefficient. This paper presents a greedy
neural word segmenter with balanced word and character embedding inputs to
alleviate the existing drawbacks. Our segmenter is truly end-to-end, capable of
performing segmentation much faster and even more accurate than
state-of-the-art neural models on Chinese benchmark datasets.Comment: To appear in ACL201
Fast Rhetorical Structure Theory Discourse Parsing
In recent years, There has been a variety of research on discourse parsing,
particularly RST discourse parsing. Most of the recent work on RST parsing has
focused on implementing new types of features or learning algorithms in order
to improve accuracy, with relatively little focus on efficiency, robustness, or
practical use. Also, most implementations are not widely available. Here, we
describe an RST segmentation and parsing system that adapts models and feature
sets from various previous work, as described below. Its accuracy is near
state-of-the-art, and it was developed to be fast, robust, and practical. For
example, it can process short documents such as news articles or essays in less
than a second
Neural Word Segmentation with Rich Pretraining
Neural word segmentation research has benefited from large-scale raw texts by
leveraging them for pretraining character and word embeddings. On the other
hand, statistical segmentation research has exploited richer sources of
external information, such as punctuation, automatic segmentation and POS. We
investigate the effectiveness of a range of external training sources for
neural word segmentation by building a modular segmentation model, pretraining
the most important submodule using rich external sources. Results show that
such pretraining significantly improves the model, leading to accuracies
competitive to the best methods on six benchmarks.Comment: Accepted by ACL 201
- …