3 research outputs found
Conductor Galloping Prediction on Imbalanced Datasets: SVM with Smart Sampling
Conductor galloping is the high-amplitude, low-frequency oscillation of
overhead power lines due to wind. Such movements may lead to severe damages to
transmission lines, and hence pose significant risks to the power system
operation. In this paper, we target to design a prediction framework for
conductor galloping. The difficulty comes from imbalanced dataset as galloping
happens rarely. By examining the impacts of data balance and data volume on the
prediction performance, we propose to employ proper sample adjustment methods
to achieve better performance. Numerical study suggests that using only three
features, together with over sampling, the SVM based prediction framework
achieves an F_1-score of 98.9%
Evaluating Word Embedding Models: Methods and Experimental Results
Extensive evaluation on a large number of word embedding models for language
processing applications is conducted in this work. First, we introduce popular
word embedding models and discuss desired properties of word models and
evaluation methods (or evaluators). Then, we categorize evaluators into
intrinsic and extrinsic two types. Intrinsic evaluators test the quality of a
representation independent of specific natural language processing tasks while
extrinsic evaluators use word embeddings as input features to a downstream task
and measure changes in performance metrics specific to that task. We report
experimental results of intrinsic and extrinsic evaluators on six word
embedding models. It is shown that different evaluators focus on different
aspects of word models, and some are more correlated with natural language
processing tasks. Finally, we adopt correlation analysis to study performance
consistency of extrinsic and intrinsic evalutors.Comment: 13 page
SG-Net: Syntax Guided Transformer for Language Representation
Understanding human language is one of the key themes of artificial
intelligence. For language representation, the capacity of effectively modeling
the linguistic knowledge from the detail-riddled and lengthy texts and getting
rid of the noises is essential to improve its performance. Traditional
attentive models attend to all words without explicit constraint, which results
in inaccurate concentration on some dispensable words. In this work, we propose
using syntax to guide the text modeling by incorporating explicit syntactic
constraints into attention mechanisms for better linguistically motivated word
representations. In detail, for self-attention network (SAN) sponsored
Transformer-based encoder, we introduce syntactic dependency of interest (SDOI)
design into the SAN to form an SDOI-SAN with syntax-guided self-attention.
Syntax-guided network (SG-Net) is then composed of this extra SDOI-SAN and the
SAN from the original Transformer encoder through a dual contextual
architecture for better linguistics inspired representation. The proposed
SG-Net is applied to typical Transformer encoders. Extensive experiments on
popular benchmark tasks, including machine reading comprehension, natural
language inference, and neural machine translation show the effectiveness of
the proposed SG-Net design.Comment: The early version accepted by IEEE Transactions on Pattern Analysis
and Machine Intelligence (TPAMI). Journal extension of arXiv:1908.05147 (AAAI
2020