43 research outputs found

    Interpretable NLG for Task-oriented Dialogue Systems with Heterogeneous Rendering Machines

    Full text link
    End-to-end neural networks have achieved promising performances in natural language generation (NLG). However, they are treated as black boxes and lack interpretability. To address this problem, we propose a novel framework, heterogeneous rendering machines (HRM), that interprets how neural generators render an input dialogue act (DA) into an utterance. HRM consists of a renderer set and a mode switcher. The renderer set contains multiple decoders that vary in both structure and functionality. For every generation step, the mode switcher selects an appropriate decoder from the renderer set to generate an item (a word or a phrase). To verify the effectiveness of our method, we have conducted extensive experiments on 5 benchmark datasets. In terms of automatic metrics (e.g., BLEU), our model is competitive with the current state-of-the-art method. The qualitative analysis shows that our model can interpret the rendering process of neural generators well. Human evaluation also confirms the interpretability of our proposed approach.Comment: Accepted as a conference paper at AAAI 202

    Zero-Shot End-to-End Spoken Language Understanding via Cross-Modal Selective Self-Training

    Full text link
    End-to-end (E2E) spoken language understanding (SLU) is constrained by the cost of collecting speech-semantics pairs, especially when label domains change. Hence, we explore \textit{zero-shot} E2E SLU, which learns E2E SLU without speech-semantics pairs, instead using only speech-text and text-semantics pairs. Previous work achieved zero-shot by pseudolabeling all speech-text transcripts with a natural language understanding (NLU) model learned on text-semantics corpora. However, this method requires the domains of speech-text and text-semantics to match, which often mismatch due to separate collections. Furthermore, using the entire collected speech-text corpus from any domains leads to \textit{imbalance} and \textit{noise} issues. To address these, we propose \textit{cross-modal selective self-training} (CMSST). CMSST tackles imbalance by clustering in a joint space of the three modalities (speech, text, and semantics) and handles label noise with a selection network. We also introduce two benchmarks for zero-shot E2E SLU, covering matched and found speech (mismatched) settings. Experiments show that CMSST improves performance in both two settings, with significantly reduced sample sizes and training time. Our code and data are released in https://github.com/amazon-science/zero-shot-E2E-slu.Comment: 18 pages, 7 figure

    Enhancing Abstractiveness of Summarization Models through Calibrated Distillation

    Full text link
    Sequence-level knowledge distillation reduces the size of Seq2Seq models for more efficient abstractive summarization. However, it often leads to a loss of abstractiveness in summarization. In this paper, we propose a novel approach named DisCal to enhance the level of abstractiveness (measured by n-gram overlap) without sacrificing the informativeness (measured by ROUGE) of generated summaries. DisCal exposes diverse pseudo summaries with two supervision to the student model. Firstly, the best pseudo summary is identified in terms of abstractiveness and informativeness and used for sequence-level distillation. Secondly, their ranks are used to ensure the student model to assign higher prediction scores to summaries with higher ranks. Our experiments show that DisCal outperforms prior methods in abstractive summarization distillation, producing highly abstractive and informative summaries.Comment: Accepted at EMNLP-Findings 202

    OmniEvent: A Comprehensive, Fair, and Easy-to-Use Toolkit for Event Understanding

    Full text link
    Event understanding aims at understanding the content and relationship of events within texts, which covers multiple complicated information extraction tasks: event detection, event argument extraction, and event relation extraction. To facilitate related research and application, we present an event understanding toolkit OmniEvent, which features three desiderata: (1) Comprehensive. OmniEvent supports mainstream modeling paradigms of all the event understanding tasks and the processing of 15 widely-used English and Chinese datasets. (2) Fair. OmniEvent carefully handles the inconspicuous evaluation pitfalls reported in Peng et al. (2023), which ensures fair comparisons between different models. (3) Easy-to-use. OmniEvent is designed to be easily used by users with varying needs. We provide off-the-shelf models that can be directly deployed as web services. The modular framework also enables users to easily implement and evaluate new event understanding models with OmniEvent. The toolkit (https://github.com/THU-KEG/OmniEvent) is publicly released along with the demonstration website and video (https://omnievent.xlore.cn/)

    Improved Data Transmission Scheme of Network Coding Based on Access Point Optimization in VANET

    Get PDF
    VANET is a hot spot of intelligent transportation researches. For vehicle users, the file sharing and content distribution through roadside access points (AP) as well as the vehicular ad hoc networks (VANET) have been an important complement to that cellular network. So the AP deployment is one of the key issues to improve the communication performance of VANET. In this paper, an access point optimization method is proposed based on particle swarm optimization algorithm. The transmission performances of the routing protocol with random linear network coding before and after the access point optimization are analyzed. The simulation results show the optimization model greatly affects the VANET transmission performances based on network coding, and it can enhance the delivery rate by 25% and 14% and reduce the average delay of transmission by 38% and 33%

    Ionic Liquid-Modulated Synthesis of Porous Worm-Like Gold with Strong SERS Response and Superior Catalytic Activities

    No full text
    Porous gold with well-defined shape and size have aroused extensive research enthusiasm due to their prominent properties in various applications. However, it is still a great challenge to explore a simple, green, and low-cost route to fabricate porous gold with a “clean” surface. In this work, porous worm-like Au has been easily synthesized in a one-step procedure from aqueous solution at room temperature under the action of ionic liquid tetrapropylammonium glycine ([N3333][Gly]). It is shown that the as-prepared porous worm-like Au has the length from 0.3 to 0.6 μm and the width of approximately 100–150 nm, and it is composed of lots of small nanoparticles about 6–12 nm in diameter. With rhodamine 6G (R6G) as a probe molecule, porous worm-like Au displays remarkable surface enhanced Raman scattering (SERS) sensitivity (detection limit is lower than 10−13 M), and extremely high reproducibility (average relative standard deviations is less than 2%). At the same time, owing to significantly high specific surface area, various pore sizes and plenty of crystal defects, porous worm-like Au also exhibits excellent catalytic performance in the reduction of nitroaromatics, such as p-nitrophenol and p-nitroaniline, which can be completely converted within only 100 s and 150 s, respectively. It is expected that the as-prepared porous worm-like Au with porous and self-supported structures will also present the encouraging advances in electrocatalysis, sensing, and many others
    corecore