43 research outputs found
Interpretable NLG for Task-oriented Dialogue Systems with Heterogeneous Rendering Machines
End-to-end neural networks have achieved promising performances in natural
language generation (NLG). However, they are treated as black boxes and lack
interpretability. To address this problem, we propose a novel framework,
heterogeneous rendering machines (HRM), that interprets how neural generators
render an input dialogue act (DA) into an utterance. HRM consists of a renderer
set and a mode switcher. The renderer set contains multiple decoders that vary
in both structure and functionality. For every generation step, the mode
switcher selects an appropriate decoder from the renderer set to generate an
item (a word or a phrase). To verify the effectiveness of our method, we have
conducted extensive experiments on 5 benchmark datasets. In terms of automatic
metrics (e.g., BLEU), our model is competitive with the current
state-of-the-art method. The qualitative analysis shows that our model can
interpret the rendering process of neural generators well. Human evaluation
also confirms the interpretability of our proposed approach.Comment: Accepted as a conference paper at AAAI 202
Zero-Shot End-to-End Spoken Language Understanding via Cross-Modal Selective Self-Training
End-to-end (E2E) spoken language understanding (SLU) is constrained by the
cost of collecting speech-semantics pairs, especially when label domains
change. Hence, we explore \textit{zero-shot} E2E SLU, which learns E2E SLU
without speech-semantics pairs, instead using only speech-text and
text-semantics pairs. Previous work achieved zero-shot by pseudolabeling all
speech-text transcripts with a natural language understanding (NLU) model
learned on text-semantics corpora. However, this method requires the domains of
speech-text and text-semantics to match, which often mismatch due to separate
collections. Furthermore, using the entire collected speech-text corpus from
any domains leads to \textit{imbalance} and \textit{noise} issues. To address
these, we propose \textit{cross-modal selective self-training} (CMSST). CMSST
tackles imbalance by clustering in a joint space of the three modalities
(speech, text, and semantics) and handles label noise with a selection network.
We also introduce two benchmarks for zero-shot E2E SLU, covering matched and
found speech (mismatched) settings. Experiments show that CMSST improves
performance in both two settings, with significantly reduced sample sizes and
training time. Our code and data are released in
https://github.com/amazon-science/zero-shot-E2E-slu.Comment: 18 pages, 7 figure
Enhancing Abstractiveness of Summarization Models through Calibrated Distillation
Sequence-level knowledge distillation reduces the size of Seq2Seq models for
more efficient abstractive summarization. However, it often leads to a loss of
abstractiveness in summarization. In this paper, we propose a novel approach
named DisCal to enhance the level of abstractiveness (measured by n-gram
overlap) without sacrificing the informativeness (measured by ROUGE) of
generated summaries. DisCal exposes diverse pseudo summaries with two
supervision to the student model. Firstly, the best pseudo summary is
identified in terms of abstractiveness and informativeness and used for
sequence-level distillation. Secondly, their ranks are used to ensure the
student model to assign higher prediction scores to summaries with higher
ranks. Our experiments show that DisCal outperforms prior methods in
abstractive summarization distillation, producing highly abstractive and
informative summaries.Comment: Accepted at EMNLP-Findings 202
OmniEvent: A Comprehensive, Fair, and Easy-to-Use Toolkit for Event Understanding
Event understanding aims at understanding the content and relationship of
events within texts, which covers multiple complicated information extraction
tasks: event detection, event argument extraction, and event relation
extraction. To facilitate related research and application, we present an event
understanding toolkit OmniEvent, which features three desiderata: (1)
Comprehensive. OmniEvent supports mainstream modeling paradigms of all the
event understanding tasks and the processing of 15 widely-used English and
Chinese datasets. (2) Fair. OmniEvent carefully handles the inconspicuous
evaluation pitfalls reported in Peng et al. (2023), which ensures fair
comparisons between different models. (3) Easy-to-use. OmniEvent is designed to
be easily used by users with varying needs. We provide off-the-shelf models
that can be directly deployed as web services. The modular framework also
enables users to easily implement and evaluate new event understanding models
with OmniEvent. The toolkit (https://github.com/THU-KEG/OmniEvent) is publicly
released along with the demonstration website and video
(https://omnievent.xlore.cn/)
Improved Data Transmission Scheme of Network Coding Based on Access Point Optimization in VANET
VANET is a hot spot of intelligent transportation researches. For vehicle users, the file sharing and content distribution through roadside access points (AP) as well as the vehicular ad hoc networks (VANET) have been an important complement to that cellular network. So the AP deployment is one of the key issues to improve the communication performance of VANET. In this paper, an access point optimization method is proposed based on particle swarm optimization algorithm. The transmission performances of the routing protocol with random linear network coding before and after the access point optimization are analyzed. The simulation results show the optimization model greatly affects the VANET transmission performances based on network coding, and it can enhance the delivery rate by 25% and 14% and reduce the average delay of transmission by 38% and 33%
Ionic Liquid-Modulated Synthesis of Porous Worm-Like Gold with Strong SERS Response and Superior Catalytic Activities
Porous gold with well-defined shape and size have aroused extensive research enthusiasm due to their prominent properties in various applications. However, it is still a great challenge to explore a simple, green, and low-cost route to fabricate porous gold with a “clean” surface. In this work, porous worm-like Au has been easily synthesized in a one-step procedure from aqueous solution at room temperature under the action of ionic liquid tetrapropylammonium glycine ([N3333][Gly]). It is shown that the as-prepared porous worm-like Au has the length from 0.3 to 0.6 μm and the width of approximately 100–150 nm, and it is composed of lots of small nanoparticles about 6–12 nm in diameter. With rhodamine 6G (R6G) as a probe molecule, porous worm-like Au displays remarkable surface enhanced Raman scattering (SERS) sensitivity (detection limit is lower than 10−13 M), and extremely high reproducibility (average relative standard deviations is less than 2%). At the same time, owing to significantly high specific surface area, various pore sizes and plenty of crystal defects, porous worm-like Au also exhibits excellent catalytic performance in the reduction of nitroaromatics, such as p-nitrophenol and p-nitroaniline, which can be completely converted within only 100 s and 150 s, respectively. It is expected that the as-prepared porous worm-like Au with porous and self-supported structures will also present the encouraging advances in electrocatalysis, sensing, and many others