209 research outputs found
Improving Medical Dialogue Generation with Abstract Meaning Representations
Medical Dialogue Generation serves a critical role in telemedicine by
facilitating the dissemination of medical expertise to patients. Existing
studies focus on incorporating textual representations, which have limited
their ability to represent the semantics of text, such as ignoring important
medical entities. To enhance the model's understanding of the textual semantics
and the medical knowledge including entities and relations, we introduce the
use of Abstract Meaning Representations (AMR) to construct graphical
representations that delineate the roles of language constituents and medical
entities within the dialogues. In this paper, We propose a novel framework that
models dialogues between patients and healthcare professionals using AMR
graphs, where the neural networks incorporate textual and graphical knowledge
with a dual attention mechanism. Experimental results show that our framework
outperforms strong baseline models in medical dialogue generation,
demonstrating the effectiveness of AMR graphs in enhancing the representations
of medical knowledge and logical relationships. Furthermore, to support future
research in this domain, we provide the corresponding source code at
https://github.com/Bernard-Yang/MedDiaAMR.Comment: Submitted to ICASSP 202
Improving Medical Dialogue Generation with Abstract Meaning Representations
Medical Dialogue Generation serves a critical role in telemedicine by facilitating the dissemination of medical expertise to patients. Existing studies focus on incorporating textual representations, which have limited their ability to represent the semantics of text, such as ignoring important medical entities. To enhance the model's understanding of the textual semantics and the medical knowledge including entities and relations, we introduce the use of Abstract Meaning Representations (AMR) to construct graphical representations that delineate the roles of language constituents and medical entities within the dialogues. In this paper, We propose a novel framework that models dialogues between patients and healthcare professionals using AMR graphs, where the neural networks incorporate textual and graphical knowledge with a dual attention mechanism. Experimental results show that our framework outperforms strong baseline models in medical dialogue generation, demonstrating the effectiveness of AMR graphs in enhancing the representations of medical knowledge and logical relationships. Furthermore, to support future research in this domain, we provide the corresponding source code at https://github.com/Bernard-Yang/MedDiaAMR
Build A Multi-Dimensional Collaborative Logistics Talent Training Model -Taking Chengdu University of Information Technology as an example
The “Opinions of the Ministry of Education on Accelerating the Construction of High-level Undergraduate Education to Comprehensively Improve the Ability of Talent Training (Jiao Gao [2018] No. 2)” clearly requires that the formation of a high-level undergraduate talent training system should be accelerated around the overall improvement of talent training capabilities. After 18 years of development, the logistics major of our school has made certain achievements in professional construction, teacher team construction, teacher teaching ability construction and experimental teaching guarantee construction, forming a relatively complete talent training system, but the social adaptation of talent training Sex still needs to be improved. This paper adheres to the problem orientation, based on the concept of collaborative education, and builds a multi-dimensional collaborative education talent training model from multiple perspectives, such as promoting the integration of production and education, condensing professional characteristics, implementing segmented training and building a capacity improvement system, and comprehensively improving the ability of logistics personnel training
Non uniform shrinkages of double-walled carbon nanotube as induced by electron beam irradiation
Electron beam-induced nanoinstabilities of pristine double-walled carbon nanotubes (DWCNTs) of two different configurations, one fixed at both ends and another fixed at only one end, were in-situ investigated in transmission electron microscope at room temperature. It was observed that the DWCNT fixed at both ends shrank in its diameter uniformly. Meanwhile, the DWCNT fixed at only one end intriguingly shrank preferentially from its free cap end along its axial direction whereas its diameter shrinkage was offset. A mechanism of "diffusion" along with "evaporation" at room temperature which is driven by the nanocurvature of the DWCNTs, and the athermal activation induced by the electron beam was proposed to elucidate the observed phenomena. The effect of the interlayer interaction of the DWCNTs was also discussed
Effective Distillation of Table-based Reasoning Ability from LLMs
Large Language Models (LLMs) have demonstrated remarkable performance across a wide range of natural language processing tasks. However, their enormous parameter size and extremely high requirements for compute power pose challenges for their practical deployment. Recent research has revealed that specific capabilities of LLMs, such as numerical reasoning, can be transferred to smaller models through distillation. Some studies explore the potential of leveraging LLMs to perform table-based reasoning. However, there has been no prior work focusing on table reasoning skills in smaller models specifically tailored for scientific table-to-text generation tasks. In this paper, we propose a novel table-based reasoning distillation approach, with the aim of distilling LLMs into tailored smaller models. Our experimental results have shown that a 220 million parameter model (Flan-T5-base) fine-tuned using distilled data, not only achieves a significant improvement compared to traditionally fine-tuned baselines, but also surpasses specific LLMs on a scientific table-to-text generation dataset. Our code is available at https://github.com/Bernard-Yang/DistillTableCoT
Effective Distillation of Table-based Reasoning Ability from LLMs
Large Language Models (LLMs) have demonstrated remarkable performance across
a wide range of natural language processing tasks. However, their remarkable
parameter size and their impressive high requirement of computing resources
pose challenges for their practical deployment. Recent research has revealed
that specific capabilities of LLMs, such as numerical reasoning, can be
transferred to smaller models through distillation. Some studies explore the
potential of leveraging LLMs to perform table-based reasoning. Nevertheless,
prior to our work, there has been no investigation into the prospect of
specialising table reasoning skills in smaller models specifically tailored for
table-to-text generation tasks. In this paper, we propose a novel table-based
reasoning distillation, with the aim of distilling distilling LLMs into
tailored, smaller models specifically designed for table-based reasoning task.
Experimental results have shown that a 0.22 billion parameter model
(Flan-T5-base) fine-tuned using distilled data, not only achieves a significant
improvement compared to traditionally fine-tuned baselines but also surpasses
specific LLMs like gpt-3.5-turbo on the scientific table-to-text generation
dataset (SciGen). The code and data are released in
https://github.com/Bernard-Yang/TableDistill
Evaluating Open-Domain Dialogues in Latent Space with Next Sentence Prediction and Mutual Information
The long-standing one-to-many issue of the open-domain dialogues poses
significant challenges for automatic evaluation methods, i.e., there may be
multiple suitable responses which differ in semantics for a given
conversational context. To tackle this challenge, we propose a novel
learning-based automatic evaluation metric (CMN), which can robustly evaluate
open-domain dialogues by augmenting Conditional Variational Autoencoders
(CVAEs) with a Next Sentence Prediction (NSP) objective and employing Mutual
Information (MI) to model the semantic similarity of text in the latent space.
Experimental results on two open-domain dialogue datasets demonstrate the
superiority of our method compared with a wide range of baselines, especially
in handling responses which are distant to the golden reference responses in
semantics.Comment: Accepted at ACL202
Solutions to a discrete resonance problem with eigenparameter-dependent boundary conditions
In this paper, we considered the existence of solutions to a discrete second-order resonance problem with eigenparameter-dependent boundary conditions. We first transformed the resonance problem into its corresponding equivalent system using the Lyapunov-Schmidt method. In addition, using Schauder's fixed-point theorem and the connectivity theories of the solution set of compact vector fields, we obtained the existence and multiplicity of solutions to the second-order discrete resonance problem with eigenparameter-dependent boundary conditions
Audio Contrastive based Fine-tuning
Audio classification plays a crucial role in speech and sound processing
tasks with a wide range of applications. There still remains a challenge of
striking the right balance between fitting the model to the training data
(avoiding overfitting) and enabling it to generalise well to a new domain.
Leveraging the transferability of contrastive learning, we introduce Audio
Contrastive-based Fine-tuning (AudioConFit), an efficient approach
characterised by robust generalisability. Empirical experiments on a variety of
audio classification tasks demonstrate the effectiveness and robustness of our
approach, which achieves state-of-the-art results in various settings.Comment: Under revie
- …