209 research outputs found

    Improving Medical Dialogue Generation with Abstract Meaning Representations

    Full text link
    Medical Dialogue Generation serves a critical role in telemedicine by facilitating the dissemination of medical expertise to patients. Existing studies focus on incorporating textual representations, which have limited their ability to represent the semantics of text, such as ignoring important medical entities. To enhance the model's understanding of the textual semantics and the medical knowledge including entities and relations, we introduce the use of Abstract Meaning Representations (AMR) to construct graphical representations that delineate the roles of language constituents and medical entities within the dialogues. In this paper, We propose a novel framework that models dialogues between patients and healthcare professionals using AMR graphs, where the neural networks incorporate textual and graphical knowledge with a dual attention mechanism. Experimental results show that our framework outperforms strong baseline models in medical dialogue generation, demonstrating the effectiveness of AMR graphs in enhancing the representations of medical knowledge and logical relationships. Furthermore, to support future research in this domain, we provide the corresponding source code at https://github.com/Bernard-Yang/MedDiaAMR.Comment: Submitted to ICASSP 202

    Improving Medical Dialogue Generation with Abstract Meaning Representations

    Get PDF
    Medical Dialogue Generation serves a critical role in telemedicine by facilitating the dissemination of medical expertise to patients. Existing studies focus on incorporating textual representations, which have limited their ability to represent the semantics of text, such as ignoring important medical entities. To enhance the model's understanding of the textual semantics and the medical knowledge including entities and relations, we introduce the use of Abstract Meaning Representations (AMR) to construct graphical representations that delineate the roles of language constituents and medical entities within the dialogues. In this paper, We propose a novel framework that models dialogues between patients and healthcare professionals using AMR graphs, where the neural networks incorporate textual and graphical knowledge with a dual attention mechanism. Experimental results show that our framework outperforms strong baseline models in medical dialogue generation, demonstrating the effectiveness of AMR graphs in enhancing the representations of medical knowledge and logical relationships. Furthermore, to support future research in this domain, we provide the corresponding source code at https://github.com/Bernard-Yang/MedDiaAMR

    Build A Multi-Dimensional Collaborative Logistics Talent Training Model -Taking Chengdu University of Information Technology as an example

    Get PDF
    The “Opinions of the Ministry of Education on Accelerating the Construction of High-level Undergraduate Education to Comprehensively Improve the Ability of Talent Training (Jiao Gao [2018] No. 2)” clearly requires that the formation of a high-level undergraduate talent training system should be accelerated around the overall improvement of talent training capabilities. After 18 years of development, the logistics major of our school has made certain achievements in professional construction, teacher team construction, teacher teaching ability construction and experimental teaching guarantee construction, forming a relatively complete talent training system, but the social adaptation of talent training Sex still needs to be improved. This paper adheres to the problem orientation, based on the concept of collaborative education, and builds a multi-dimensional collaborative education talent training model from multiple perspectives, such as promoting the integration of production and education, condensing professional characteristics, implementing segmented training and building a capacity improvement system, and comprehensively improving the ability of logistics personnel training

    Non uniform shrinkages of double-walled carbon nanotube as induced by electron beam irradiation

    Get PDF
    Electron beam-induced nanoinstabilities of pristine double-walled carbon nanotubes (DWCNTs) of two different configurations, one fixed at both ends and another fixed at only one end, were in-situ investigated in transmission electron microscope at room temperature. It was observed that the DWCNT fixed at both ends shrank in its diameter uniformly. Meanwhile, the DWCNT fixed at only one end intriguingly shrank preferentially from its free cap end along its axial direction whereas its diameter shrinkage was offset. A mechanism of "diffusion" along with "evaporation" at room temperature which is driven by the nanocurvature of the DWCNTs, and the athermal activation induced by the electron beam was proposed to elucidate the observed phenomena. The effect of the interlayer interaction of the DWCNTs was also discussed

    Effective Distillation of Table-based Reasoning Ability from LLMs

    Get PDF
    Large Language Models (LLMs) have demonstrated remarkable performance across a wide range of natural language processing tasks. However, their enormous parameter size and extremely high requirements for compute power pose challenges for their practical deployment. Recent research has revealed that specific capabilities of LLMs, such as numerical reasoning, can be transferred to smaller models through distillation. Some studies explore the potential of leveraging LLMs to perform table-based reasoning. However, there has been no prior work focusing on table reasoning skills in smaller models specifically tailored for scientific table-to-text generation tasks. In this paper, we propose a novel table-based reasoning distillation approach, with the aim of distilling LLMs into tailored smaller models. Our experimental results have shown that a 220 million parameter model (Flan-T5-base) fine-tuned using distilled data, not only achieves a significant improvement compared to traditionally fine-tuned baselines, but also surpasses specific LLMs on a scientific table-to-text generation dataset. Our code is available at https://github.com/Bernard-Yang/DistillTableCoT

    Effective Distillation of Table-based Reasoning Ability from LLMs

    Full text link
    Large Language Models (LLMs) have demonstrated remarkable performance across a wide range of natural language processing tasks. However, their remarkable parameter size and their impressive high requirement of computing resources pose challenges for their practical deployment. Recent research has revealed that specific capabilities of LLMs, such as numerical reasoning, can be transferred to smaller models through distillation. Some studies explore the potential of leveraging LLMs to perform table-based reasoning. Nevertheless, prior to our work, there has been no investigation into the prospect of specialising table reasoning skills in smaller models specifically tailored for table-to-text generation tasks. In this paper, we propose a novel table-based reasoning distillation, with the aim of distilling distilling LLMs into tailored, smaller models specifically designed for table-based reasoning task. Experimental results have shown that a 0.22 billion parameter model (Flan-T5-base) fine-tuned using distilled data, not only achieves a significant improvement compared to traditionally fine-tuned baselines but also surpasses specific LLMs like gpt-3.5-turbo on the scientific table-to-text generation dataset (SciGen). The code and data are released in https://github.com/Bernard-Yang/TableDistill

    Evaluating Open-Domain Dialogues in Latent Space with Next Sentence Prediction and Mutual Information

    Full text link
    The long-standing one-to-many issue of the open-domain dialogues poses significant challenges for automatic evaluation methods, i.e., there may be multiple suitable responses which differ in semantics for a given conversational context. To tackle this challenge, we propose a novel learning-based automatic evaluation metric (CMN), which can robustly evaluate open-domain dialogues by augmenting Conditional Variational Autoencoders (CVAEs) with a Next Sentence Prediction (NSP) objective and employing Mutual Information (MI) to model the semantic similarity of text in the latent space. Experimental results on two open-domain dialogue datasets demonstrate the superiority of our method compared with a wide range of baselines, especially in handling responses which are distant to the golden reference responses in semantics.Comment: Accepted at ACL202

    Solutions to a discrete resonance problem with eigenparameter-dependent boundary conditions

    Get PDF
    In this paper, we considered the existence of solutions to a discrete second-order resonance problem with eigenparameter-dependent boundary conditions. We first transformed the resonance problem into its corresponding equivalent system using the Lyapunov-Schmidt method. In addition, using Schauder's fixed-point theorem and the connectivity theories of the solution set of compact vector fields, we obtained the existence and multiplicity of solutions to the second-order discrete resonance problem with eigenparameter-dependent boundary conditions

    Audio Contrastive based Fine-tuning

    Full text link
    Audio classification plays a crucial role in speech and sound processing tasks with a wide range of applications. There still remains a challenge of striking the right balance between fitting the model to the training data (avoiding overfitting) and enabling it to generalise well to a new domain. Leveraging the transferability of contrastive learning, we introduce Audio Contrastive-based Fine-tuning (AudioConFit), an efficient approach characterised by robust generalisability. Empirical experiments on a variety of audio classification tasks demonstrate the effectiveness and robustness of our approach, which achieves state-of-the-art results in various settings.Comment: Under revie
    corecore