26,111 research outputs found

    A Controllable Model of Grounded Response Generation

    Full text link
    Current end-to-end neural conversation models inherently lack the flexibility to impose semantic control in the response generation process, often resulting in uninteresting responses. Attempts to boost informativeness alone come at the expense of factual accuracy, as attested by pretrained language models' propensity to "hallucinate" facts. While this may be mitigated by access to background knowledge, there is scant guarantee of relevance and informativeness in generated responses. We propose a framework that we call controllable grounded response generation (CGRG), in which lexical control phrases are either provided by a user or automatically extracted by a control phrase predictor from dialogue context and grounding knowledge. Quantitative and qualitative results show that, using this framework, a transformer based model with a novel inductive attention mechanism, trained on a conversation-like Reddit dataset, outperforms strong generation baselines.Comment: AAAI 202

    지식 기반 λŒ€ν™”μ—μ„œμ˜ λŒ€ν™” νŠΉμ„±μ„ ν™œμš©ν•œ 지식 선택 및 λž­ν‚Ή 방법

    Get PDF
    ν•™μœ„λ…Όλ¬Έ(박사) -- μ„œμšΈλŒ€ν•™κ΅λŒ€ν•™μ› : κ³΅κ³ΌλŒ€ν•™ 전기·컴퓨터곡학뢀, 2022. 8. 이상ꡬ.Knowledge grounded conversation (KGC) model aims to generate informative responses relevant to both conversation history and external knowledge. One of the most important parts of KGC models is to find the knowledge which provides the basis on which the responses are grounded. If the model selects inappropriate knowledge, it may produce responses that are irrelevant or lack knowledge. In this dissertation, we study the methods of leveraging conversational characteristics to select or rank the knowledge for knowledge grounded conversation. In particular, this dissertation provides novel two methods, where one of which focuses on the sequential structure of multi-turn conversation, and the other focuses on utilizing local context and topic of a long conversation. We first propose two knowledge selection strategies of which one preserves the sequential matching features and the other encodes the sequential nature of the conversation. Second, we propose a novel knowledge ranking model that composes an appropriate range of relevant documents by exploiting both the topic keywords and local context of a conversation. In addition, we apply the knowledge ranking model in quote recommendation with our new quote recommendation framework that provides hard negative samples to the model. Our experimental results show that the KGC models based on our proposed knowledge selection and ranking methods outperform the competitive models in terms of groundness and relevance.지식 기반 λŒ€ν™” λͺ¨λΈμ€ λŒ€ν™” 기둝과 μ™ΈλΆ€ 지식 이 두 가지 λͺ¨λ‘μ— κ΄€λ ¨λœ 응닡을 μƒμ„±ν•˜λŠ” 것을 λͺ©ν‘œλ‘œ ν•œλ‹€. 지식 기반 λŒ€ν™” λͺ¨λΈμ˜ κ°€μž₯ μ€‘μš”ν•œ λΆ€λΆ„ 쀑 ν•˜λ‚˜λŠ” μ‘λ‹΅μ˜ κΈ°λ°˜μ„ μ œκ³΅ν•˜λŠ” 지식을 μ°ΎλŠ” 것이닀. 지식 기반 λͺ¨λΈμ΄ 주어진 λ¬Έλ§₯에 λΆ€μ ν•©ν•œ 지식을 μ°ΎλŠ” 경우 관련성이 λ–¨μ–΄μ§€κ±°λ‚˜ 지식이 λΆ€μ‘±ν•œ 응닡이 생성될 수 μžˆλ‹€. 이 문제λ₯Ό ν•΄κ²°ν•˜κΈ° μœ„ν•΄ 이 λ…Όλ¬Έμ—μ„œλŠ” 지식 기반 λŒ€ν™”λ₯Ό μœ„ν•΄ λŒ€ν™” μ—¬λŸ¬ νŠΉμ„±μ„ ν™œμš©ν•˜μ—¬ 지식을 μ„ μ •ν•˜λŠ” 지식 선택 λͺ¨λΈκ³Ό 지식 μˆœμœ„ λͺ¨λΈμ„ μ œμ‹œν•œλ‹€. ꡬ체적으둜 λ³Έ λ…Όλ¬Έμ—μ„œλŠ” 닀쀑 ν„΄ λŒ€ν™”μ—μ„œμ˜ 순차적 ꡬ쑰 λ˜λŠ” 응닡 이전 λ¬Έλ§₯κ³Ό λŒ€ν™”μ˜ 주제λ₯Ό ν™œμš©ν•˜λŠ” μƒˆλ‘œμš΄ 두 가지 방법을 μ œμ‹œν•œλ‹€. 첫 번째 λ°©λ²•μœΌλ‘œμ¨ λ³Έ 논문은 두 가지 지식 선택 μ „λž΅μ„ μ œμ•ˆν•œλ‹€. μ œμ•ˆν•˜λŠ” μ „λž΅ 쀑 ν•˜λ‚˜λŠ” 지식과 λŒ€ν™” ν„΄ κ°„μ˜ 순차적 맀칭 νŠΉμ§•μ„ λ³΄μ‘΄ν•˜λŠ” 방법이고 λ‹€λ₯Έ μ „λž΅μ€ λŒ€ν™”μ˜ 순차적 νŠΉμ„±μ„ μΈμ½”λ”©ν•˜μ—¬ 지식을 μ„ νƒν•˜λŠ” 방법이닀. 두 번째둜 λ³Έ 논문은 λŒ€ν™”μ˜ 주제 ν‚€μ›Œλ“œμ™€ 응닡 λ°”λ‘œ μ΄μ „μ˜ λ¬Έλ§₯을 λͺ¨λ‘ ν™œμš©ν•˜μ—¬ μ μ ˆν•œ λ²”μœ„μ˜ κ΄€λ ¨ λ¬Έμ„œλ“€λ‘œ 검색 κ²°κ³Όλ₯Ό κ΅¬μ„±ν•˜λŠ” μƒˆλ‘œμš΄ 지식 μˆœμœ„ λͺ¨λΈμ„ μ œμ•ˆν•œλ‹€. λ§ˆμ§€λ§‰μœΌλ‘œ 지식 μˆœμœ„ λͺ¨λΈμ˜ 적응성 검증을 μœ„ν•΄ μ •λ‹΅ μΈμš©κ΅¬μ™€ 의미적으둜 μœ μ‚¬ν•˜μ§€λ§Œ 정닡은 μ•„λ‹Œ 인용ꡬ의 집합을 인용ꡬ μˆœμœ„ λͺ¨λΈμ— μ œκ³΅ν•˜λŠ” 인용ꡬ μΆ”μ²œ ν”„λ ˆμž„μ›Œν¬λ₯Ό μ œμ•ˆν•œλ‹€. μ œμ•ˆλœ 지식 선택 및 μˆœμœ„ λͺ¨λΈμ„ 기반으둜 ν•˜λŠ” 지식 기반 λŒ€ν™” λͺ¨λΈμ΄ 경쟁 λͺ¨λΈλ³΄λ‹€ μ™ΈλΆ€ 지식 및 λŒ€ν™” λ¬Έλ§₯과의 κ΄€λ ¨μ„± μΈ‘λ©΄μ—μ„œ μš°μˆ˜ν•˜λ‹€λŠ” 것을 μ‚¬λžŒ κ°„μ˜ λŒ€ν™” 데이터λ₯Ό μ΄μš©ν•œ λ‹€μˆ˜μ˜ μ‹€ν—˜μ„ 톡해 κ²€μ¦ν•˜μ˜€λ‹€.Abstract 1 1. Introduction 17 2. Background and Related Works 25 2.1 Terminology 25 2.2 Overview of Technologies for Conversational Systems 27 2.2.1 Open-domain Dialogue System 27 2.2.2 Task-oriented Dialogue System 29 2.2.3 Question Answering System 29 2.3 Components of Knowledge Grounded Conversation Model 31 2.4 Related Works 36 2.4.1 KGC datasets 36 2.4.2 Soft Selection-based KGC Model 36 2.4.3 Hard Selection-based KGC Model 37 2.4.4 Retrieval-based KGC Models 39 2.4.5 Response Generation with Knowledge Integration 39 2.4.6 Quote Recommendation 42 2.5 Evaluation Methods 44 2.6 Problem Statements 47 3. Knowledge Selection with Sequential Structure of Conversation 48 3.1 Motivation 48 3.2 Reduce-Match Strategy & Match-Reduce Strategy 49 3.2.1 Backbone architecture 51 3.2.2 Reduce-Match Strategy-based Models 52 3.2.3 Match-Reduce Strategy-based Models 56 3.3 Experiments 62 3.3.1 Experimental Setup 62 3.3.2 Experimental Results 70 3.4 Analysis 72 3.4.1 Case Study 72 3.4.2 Impact of Matching Difficulty 75 3.4.3 Impact of Length of Context 77 3.4.4 Impact of Dialogue Act of Message 78 4. Knowledge Ranking with Local Context and Topic Keywords 81 4.1 Motivation 81 4.2 Retrieval-Augmented Knowledge Grounded Conversation Model 85 4.2.1 Base Model 86 4.2.2 Topic-aware Dual Matching for Knowledge Re-ranking 86 4.2.3 Data Weighting Scheme for Retrieval Augmented Generation Models 89 4.3 Experiments 90 4.3.1 Experimental Setup 90 4.3.2 Experimental Results 94 4.4 Analysis 98 4.4.1 Case Study 98 4.4.2 Ablation Study 99 4.4.3 Model Variations 104 4.4.4 Error Analysis 105 5. Application: Quote Recommendation with Knowledge Ranking 110 5.1 Motivation 110 5.2 CAGAR: A Framework for Quote Recommendation 112 5.2.1 Conversation Encoder 114 5.2.2 Quote Encoder 114 5.2.3 Candidate Generator 115 5.2.4 Re-ranker 116 5.2.5 Training and Inference 116 5.3 Experiments 117 5.3.1 Experimental Setup 117 5.3.2 Experimental Results 119 5.4 Analysis 120 5.4.1 Ablation Study 120 5.4.2 Case Study 121 5.4.3 Impact of Length of Context 121 5.4.4 Impact of Training Set Size per Quote 123 6. Conclusion 125 6.1 Contributions and Limitations 126 6.2 Future Works 128 Appendix A. Preliminary Experiments for Quote Recommendations 131 A.1 Methods 131 A.1.1 Matching Granularity Adjustment 131 A.1.2 Random Forest 133 A.1.3 Convolutional Neural Network 133 A.1.4 Recurrent Neural Network 134 A.2 Experiments 135 A.2.1 Baselines and Implementation Details 135 A.2.2 Datasets 136 A.2.3 Results and Discussions 137 초둝 162λ°•
    • …
    corecore