4,828 research outputs found

    ATP: A holistic attention integrated approach to enhance ABSA

    Full text link
    Aspect based sentiment analysis (ABSA) deals with the identification of the sentiment polarity of a review sentence towards a given aspect. Deep Learning sequential models like RNN, LSTM, and GRU are current state-of-the-art methods for inferring the sentiment polarity. These methods work well to capture the contextual relationship between the words of a review sentence. However, these methods are insignificant in capturing long-term dependencies. Attention mechanism plays a significant role by focusing only on the most crucial part of the sentence. In the case of ABSA, aspect position plays a vital role. Words near to aspect contribute more while determining the sentiment towards the aspect. Therefore, we propose a method that captures the position based information using dependency parsing tree and helps attention mechanism. Using this type of position information over a simple word-distance-based position enhances the deep learning model's performance. We performed the experiments on SemEval'14 dataset to demonstrate the effect of dependency parsing relation-based attention for ABSA

    토큰 λ‹¨μœ„ λΆ„λ₯˜λͺ¨λΈμ„ μœ„ν•œ μ€‘μš” 토큰 포착 및 μ‹œν€€μŠ€ 인코더 섀계 방법

    Get PDF
    ν•™μœ„λ…Όλ¬Έ(박사) -- μ„œμšΈλŒ€ν•™κ΅λŒ€ν•™μ› : κ³΅κ³ΌλŒ€ν•™ 전기·정보곡학뢀, 2022. 8. 정ꡐ민.With the development of internet, a great of volume of data have accumulated over time. Therefore, dealing long sequential data can become a core problem in web services. For example, streaming services such as YouTube, Netflx and Tictoc have used the user's viewing history sequence to recommend videos that users may like. Such systems have replaced the user's viewed video with each item or token to predict what item or token will be viewed next. These tasks have been defined as Token-Level Classification (TLC) tasks. Given the sequence of tokens, TLC identifies the labels of tokens in the required portion of this sequence. As mentioned above, TLC can be applied to various recommendation Systems. In addition, most of Natural Language Processing (NLP) tasks can also be formulated as TLC problem. For example, sentence and each word within the sentence can be expressed as token-level sequence. In particular, in the case of information extraction, it can be changed to a TLC task that distinguishes whether a specific word span in the sentence is information. The characteristics of TLC datasets are that they are very sparse and long. Therefore, it is a very important problem to extract only important information from the sequences and properly encode them. In this thesis, we propose the method to solve the two academic questions of TLC in Recommendation Systems and information extraction: 1) How to capture important tokens from the token sequence and 2) How to encode a token sequence into model. As deep neural networks (DNNs) have shown outstanding performance in various web application tasks, we design the RNN and Transformer-based model for recommendation systems, and information extractions. In this dissertation, we propose novel models that can extract important tokens for recommendation systems and information extraction systems. In recommendation systems, we design a BART-based system that can capture important portion of token sequence through self-attention mechanisms and consider both bidirectional and left-to-right directional information. In information systems, we present relation network-based models to focus important parts such as opinion target and neighbor words.μΈν„°λ„·μ˜ λ°œλ‹¬λ‘œ, λ§Žμ€ μ–‘μ˜ 데이터가 μ‹œκ°„μ΄ 지남에 따라 μΆ•μ λ˜μ—ˆλ‹€. μ΄λ‘œμΈν•΄ κΈ΄ 순차적 데이터λ₯Ό μ²˜λ¦¬ν•˜λŠ” 것은 μ›Ή μ„œλΉ„μŠ€μ˜ 핡심 λ¬Έμ œκ°€ λ˜μ—ˆλ‹€. 예λ₯Ό λ“€μ–΄, 유튜브, λ„·ν”Œλ¦­μŠ€, 틱톑과 같은 슀트리밍 μ„œλΉ„μŠ€λŠ” μ‚¬μš©μžμ˜ μ‹œμ²­ 기둝 μ‹œν€€μŠ€λ₯Ό μ‚¬μš©ν•˜μ—¬ μ‚¬μš©μžκ°€ μ’‹μ•„ν•  λ§Œν•œ λΉ„λ””μ˜€λ₯Ό μΆ”μ²œν•œλ‹€. μ΄λŸ¬ν•œ μ‹œμŠ€ν…œμ€ λ‹€μŒμ— μ–΄λ–€ ν•­λͺ©μ΄λ‚˜ 토큰을 λ³Ό 것인지λ₯Ό μ˜ˆμΈ‘ν•˜κΈ° μœ„ν•΄ μ‚¬μš©μžκ°€ λ³Έ λΉ„λ””μ˜€λ₯Ό 각 ν•­λͺ© λ˜λŠ” ν† ν°μœΌλ‘œ λŒ€μ²΄ν•˜μ—¬ μ‚¬μš©ν•  수 μžˆλ‹€. μ΄λŸ¬ν•œ μž‘μ—…μ€ 토큰 μˆ˜μ€€ λΆ„λ₯˜(TLC) μž‘μ—…μœΌλ‘œ μ •μ˜ν•œλ‹€. 토큰 μ‹œν€€μŠ€κ°€ 주어지면, TLCλŠ” 이 μ‹œν€€μŠ€μ˜ ν•„μš”ν•œ λΆ€λΆ„μ—μ„œ ν† ν°μ˜ 라벨을 μ‹λ³„ν•œλ‹€. μ΄λ ‡κ²Œμ™€ 같이, TLCλŠ” λ‹€μ–‘ν•œ μΆ”μ²œ μ‹œμŠ€ν…œμ— 적용될 수 μžˆλ‹€. λ˜ν•œ λŒ€λΆ€λΆ„μ˜ μžμ—°μ–΄ 처리(NLP) μž‘μ—…μ€ TLC 문제둜 곡식화될 수 μžˆλ‹€. 예λ₯Ό λ“€μ–΄, λ¬Έμž₯κ³Ό λ¬Έμž₯ λ‚΄μ˜ 각 λ‹¨μ–΄λŠ” 토큰 레벨 μ‹œν€€μŠ€λ‘œ ν‘œν˜„λ  수 μžˆλ‹€. 특히 정보 μΆ”μΆœμ˜ 경우 λ¬Έμž₯의 νŠΉμ • 단어 간격이 정보인지 μ—¬λΆ€λ₯Ό κ΅¬λΆ„ν•˜λŠ” TLC μž‘μ—…μœΌλ‘œ λ°”λ€” 수 μžˆλ‹€. TLC 데이터 μ„ΈνŠΈμ˜ νŠΉμ§•μ€ 맀우 희박(Sparse)ν•˜κ³  κΈΈλ‹€λŠ” 것이닀. λ”°λΌμ„œ μ‹œν€€μŠ€μ—μ„œ μ€‘μš”ν•œ μ •λ³΄λ§Œ μΆ”μΆœν•˜μ—¬ 적절히 μΈμ½”λ”©ν•˜λŠ” 것은 맀우 μ€‘μš”ν•œ λ¬Έμ œμ΄λ‹€. λ³Έ λ…Όλ¬Έμ—μ„œλŠ” ꢌμž₯ μ‹œμŠ€ν…œκ³Ό 정보 μΆ”μΆœμ—μ„œ TLC의 두 가지 학문적 질문- 1) 토큰 μ‹œν€€μŠ€μ—μ„œ μ€‘μš”ν•œ 토큰을 μΊ‘μ²˜ν•˜λŠ” 방법 및 2) 토큰 μ‹œν€€μŠ€λ₯Ό λͺ¨λΈλ‘œ μΈμ½”λ”©ν•˜λŠ” 방법 을 ν•΄κ²°ν•˜λŠ” 방법을 μ œμ•ˆν•œλ‹€. 심측 신경망(DNN)이 λ‹€μ–‘ν•œ μ›Ή μ• ν”Œλ¦¬μΌ€μ΄μ…˜ μž‘μ—…μ—μ„œ λ›°μ–΄λ‚œ μ„±λŠ₯을 보여 μ™”κΈ° λ•Œλ¬Έμ— μΆ”μ²œ μ‹œμŠ€ν…œ 및 정보 μΆ”μΆœμ„ μœ„ν•œ RNN 및 트랜슀포머 기반 λͺ¨λΈμ„ μ„€κ³„ν•œλ‹€. λ¨Όμ € μš°λ¦¬λŠ” 자기 주의 λ©”μ»€λ‹ˆμ¦˜μ„ 톡해 토큰 μ‹œν€€μŠ€μ˜ μ€‘μš”ν•œ 뢀뢄을 ν¬μ°©ν•˜κ³  μ–‘λ°©ν–₯ 및 쒌우 λ°©ν–₯ 정보λ₯Ό λͺ¨λ‘ κ³ λ €ν•  수 μžˆλŠ” BART 기반 μΆ”μ²œ μ‹œμŠ€ν…œμ„ μ„€κ³„ν•œλ‹€. 정보 μ‹œμŠ€ν…œμ—μ„œ, μš°λ¦¬λŠ” 의견 λŒ€μƒκ³Ό 이웃 단어와 같은 μ€‘μš”ν•œ 뢀뢄에 μ΄ˆμ μ„ λ§žμΆ”κΈ° μœ„ν•΄ 관계 λ„€νŠΈμ›Œν¬ 기반 λͺ¨λΈμ„ μ œμ‹œν•œλ‹€.1. Introduction 1 2. Token-level Classification in Recommendation Systems 8 2.1 Overview 8 2.2 Hierarchical RNN-based Recommendation Systems 19 2.3 Entangled Bidirectional Encoder to Auto-regressive Decoder for Sequential Recommendation 27 3. Token-level Classification in Information Extraction 39 3.1 Overview 39 3.2 RABERT: Relation-Aware BERT for Target-Oriented Opinion Words Extraction 49 3.3 Gated Relational Target-aware Encoder and Local Context-aware Decoder for Target-oriented Opinion Words Extraction 58 4. Conclusion 79λ°•

    Context-Guided BERT for Targeted Aspect-Based Sentiment Analysis

    Full text link
    Aspect-based sentiment analysis (ABSA) and Targeted ASBA (TABSA) allow finer-grained inferences about sentiment to be drawn from the same text, depending on context. For example, a given text can have different targets (e.g., neighborhoods) and different aspects (e.g., price or safety), with different sentiment associated with each target-aspect pair. In this paper, we investigate whether adding context to self-attention models improves performance on (T)ABSA. We propose two variants of Context-Guided BERT (CG-BERT) that learn to distribute attention under different contexts. We first adapt a context-aware Transformer to produce a CG-BERT that uses context-guided softmax-attention. Next, we propose an improved Quasi-Attention CG-BERT model that learns a compositional attention that supports subtractive attention. We train both models with pretrained BERT on two (T)ABSA datasets: SentiHood and SemEval-2014 (Task 4). Both models achieve new state-of-the-art results with our QACG-BERT model having the best performance. Furthermore, we provide analyses of the impact of context in the our proposed models. Our work provides more evidence for the utility of adding context-dependencies to pretrained self-attention-based language models for context-based natural language tasks

    Aspect-oriented Opinion Alignment Network for Aspect-Based Sentiment Classification

    Full text link
    Aspect-based sentiment classification is a crucial problem in fine-grained sentiment analysis, which aims to predict the sentiment polarity of the given aspect according to its context. Previous works have made remarkable progress in leveraging attention mechanism to extract opinion words for different aspects. However, a persistent challenge is the effective management of semantic mismatches, which stem from attention mechanisms that fall short in adequately aligning opinions words with their corresponding aspect in multi-aspect sentences. To address this issue, we propose a novel Aspect-oriented Opinion Alignment Network (AOAN) to capture the contextual association between opinion words and the corresponding aspect. Specifically, we first introduce a neighboring span enhanced module which highlights various compositions of neighboring words and given aspects. In addition, we design a multi-perspective attention mechanism that align relevant opinion information with respect to the given aspect. Extensive experiments on three benchmark datasets demonstrate that our model achieves state-of-the-art results. The source code is available at https://github.com/AONE-NLP/ABSA-AOAN.Comment: 8 pages, 5 figure, ECAI 202
    • …
    corecore