15 research outputs found

    MIRACLE: Towards Personalized Dialogue Generation with Latent-Space Multiple Personal Attribute Control

    Full text link
    Personalized dialogue systems aim to endow the chatbot agent with more anthropomorphic traits for human-like interactions. Previous approaches have explored explicitly user profile modeling using text descriptions, implicit derivation of user embeddings, or utilizing handicraft prompts for ChatGPT-like models. However, textual personas are limited in describing multi-faceted attributes (\emph{e.g.}, \emph{language style, inner character nuances}), implicit embedding suffers from personality sparsity, and handicraft prompts lack fine-grained and stable controllability. Hence, these approaches may struggle with complex personalized dialogue generation tasks that require generating controllable responses with multiple personal attributes. To this end, we propose \textbf{\textsc{Miracle}}, a novel personalized dialogue generation method through \textbf{M}ult\textbf{I}ple Pe\textbf{R}sonal \textbf{A}ttributes \textbf{C}ontrol within \textbf{L}atent-Space \textbf{E}nergy-based Models. ttributes \textbf{C}ontrol within \textbf{L}atent-Space \textbf{E}nergy-based Models. Specifically, our approach first disentangles complex personality into multi-faceted attributes. Subsequently, we employ a conditional variational auto-encoder to align with the dense personalized responses within a latent joint attribute space. We have also tailored a dedicated energy function and customized the ordinary differential equations sampling method to offer flexible attribute composition and precise attribute control. Extensive experiments demonstrate that \textsc{Miracle} outperforms several strong baselines in terms of personality controllability and response generation quality. Our dataset and code are available at \url{https://github.com/LZY-the-boys/MIRACLE}Comment: Accepted by EMNLP2023 finding

    Target Guided Emotion Aware Chat Machine

    Full text link
    The consistency of a response to a given post at semantic-level and emotional-level is essential for a dialogue system to deliver human-like interactions. However, this challenge is not well addressed in the literature, since most of the approaches neglect the emotional information conveyed by a post while generating responses. This article addresses this problem by proposing a unifed end-to-end neural architecture, which is capable of simultaneously encoding the semantics and the emotions in a post and leverage target information for generating more intelligent responses with appropriately expressed emotions. Extensive experiments on real-world data demonstrate that the proposed method outperforms the state-of-the-art methods in terms of both content coherence and emotion appropriateness.Comment: To appear on TOIS 202

    Simvastatin-Loaded Nanomicelles Enhance the Osteogenic Effect of Simvastatin

    No full text
    Objectives. The present study intended to further verify that simvastatin-loaded nanomicelles (SVNs) enhanced the role of simvastatin (SV) in promoting osteoblast differentiation in vitro and to evaluate the effect of SVNs on bone defect repair in vivo. Methods. SVNs were synthesized by dialysis. MG63 cells were subjected to intervention with 0.25 μmol/l of SVNs and SV. A 3-(4,5-dimethylthiazol-2-yl)-5-(3-carboxymethoxyphenyl)-2-(4-sulfophenyl)-2H-tetrazolium (MTS) assay kit and flow cytometry were used to determine cell proliferation activity, cell cycle distribution, and apoptosis. The osteoblastic differentiation of MG 63 cells was evaluated by measuring alkaline phosphatase (ALP) activity, ALP staining, and the expression levels of the osterix (Osx) and osteocalcin (OC) proteins. In addition, 0.5 mg of SVNs or SV was applied to the skull defect area of rabbits. Micro-CT, hematoxylin and eosin (HE) staining, and Masson’s trichrome staining were used for qualitative and quantitative evaluation of new bone in three dimensions and two dimensions. Results. The SVNs had a mean diameter of 38.97 nm. The encapsulation and drug-loading efficiencies were 54.57±3.15% and 10.91±0.63%, respectively. In vitro, SVNs and SV can inhibit the proliferation activity and promote osteogenic differentiation of MG63 cells by arresting MG63 cells at the G0/G1 phase without increasing the apoptosis rate. In vivo quantitative results showed that the bone mineral density (BMD), bone volume (BV)/total volume (TV) ratio, and trabecular number (Tb.N) in the gelatin sponge with SVNs (SVNs-GS) group and gelatin sponge with SV (SV-GS) group were 362.1%, 292.0%; 181.3%, 158.0%; and 215.2%, 181.8% of those in the blank control (BC) group, respectively. Histological results identified the new bone tissue in each group as irregular fibrous bone, and the arrangement of trabecular bone was disordered. There were significantly more osteoblasts and new capillaries around the trabecular bone in the SVNs-GS group and SV-GS group than in both the BC and drug-free nanomicelle (DFNs) groups. Both in vitro and in vivo, SVNs exhibited greater osteogenic efficacy than SV. Conclusion. SVNs significantly improved the osteogenic efficacy of SV

    OneRel: Joint Entity and Relation Extraction with One Module in One Step

    No full text
    Joint entity and relation extraction is an essential task in natural language processing and knowledge graph construction. Existing approaches usually decompose the joint extraction task into several basic modules or processing steps to make it easy to conduct. However, such a paradigm ignores the fact that the three elements of a triple are interdependent and indivisible. Therefore, previous joint methods suffer from the problems of cascading errors and redundant information. To address these issues, in this paper, we propose a novel joint entity and relation extraction model, named OneRel, which casts joint extraction as a fine-grained triple classification problem. Specifically, our model consists of a scoring-based classifier and a relation-specific horns tagging strategy. The former evaluates whether a token pair and a relation belong to a factual triple. The latter ensures a simple but effective decoding process. Extensive experimental results on two widely used datasets demonstrate that the proposed method performs better than the state-of-the-art baselines, and delivers consistent performance gain on complex scenarios of various overlapping patterns and multiple triples

    Stack-VS : stacked visual-semantic attention for image caption generation

    No full text
    Recently, automatic image caption generation has been an important focus of the work on multimodal translation task. Existing approaches can be roughly categorized into two classes, top-down and bottom-up, the former transfers the image information (called as visual-level feature) directly into a caption, and the later uses the extracted words (called as semantic-level attribute) to generate a description. However, previous methods either are typically based one-stage decoder or partially utilize part of visual-level or semantic-level information for image caption generation. In this paper, we address the problem and propose an innovative multi-stage architecture (called as Stack-VS) for rich fine-grained image caption generation, via combining bottom-up and top-down attention models to effectively handle both visual-level and semantic-level information of an input image. Specifically, we also propose a novel well-designed stack decoder model, which is constituted by a sequence of decoder cells, each of which contains two LSTM-layers work interactively to re-optimize attention weights on both visual-level feature vectors and semantic-level attribute embeddings for generating a fine-grained image caption. Extensive experiments on the popular benchmark dataset MSCOCO show the significant improvements on different evaluation metrics, i.e., the improvements on BLEU-4 / CIDEr / SPICE scores are 0.372, 1.226 and 0.216, respectively, as compared to the state-of-the-art.Published versio
    corecore