220 research outputs found

    Learning Cross-modal Context Graph for Visual Grounding

    Full text link
    Visual grounding is a ubiquitous building block in many vision-language tasks and yet remains challenging due to large variations in visual and linguistic features of grounding entities, strong context effect and the resulting semantic ambiguities. Prior works typically focus on learning representations of individual phrases with limited context information. To address their limitations, this paper proposes a language-guided graph representation to capture the global context of grounding entities and their relations, and develop a cross-modal graph matching strategy for the multiple-phrase visual grounding task. In particular, we introduce a modular graph neural network to compute context-aware representations of phrases and object proposals respectively via message propagation, followed by a graph-based matching module to generate globally consistent localization of grounding phrases. We train the entire graph neural network jointly in a two-stage strategy and evaluate it on the Flickr30K Entities benchmark. Extensive experiments show that our method outperforms the prior state of the arts by a sizable margin, evidencing the efficacy of our grounding framework. Code is available at "https://github.com/youngfly11/LCMCG-PyTorch".Comment: AAAI-202

    Automatic Generation of Electronic Medical Record Based on GPT2 Model

    Get PDF
    Writing Electronic Medical Records (EMR) as one of daily major tasks of doctors, consumes a lot of time and effort from doctors. This paper reports our efforts to generate electronic medical records using the language model. Through the training of massive real-world EMR data, the CMedGPT2 model provided by us can achieve the ideal Chinese electronic medical record generation. The experimental results prove that the generated electronic medical record text can be applied to the auxiliary medical record work to reduce the burden on the compose and provide a fast and accurate reference for composing work

    An Word2vec based on Chinese Medical Knowledge

    Get PDF
    Introducing a large amount of external prior domain knowledge will effectively improve the performance of the word embedded language model in downstream NLP tasks. Based on this assumption, we collect and collate a medical corpus data with about 36M (Million) characters and use the data of CCKS2019 as the test set to carry out multiple classifications and named entity recognition (NER) tasks with the generated word and character vectors. Compared with the results of BERT, our models obtained the ideal performance and efficiency results

    Disease Diagnosis Prediction of EMR Based on BiGRL-Att-CapsNetwork Model

    Get PDF
    Electronic Medical Records (EMR) carry a large number of diseases characteristics, history and other specific details of patients, which has great value for medical diagnosis. These data with diagnostic labels can help automated diagnostic assistant to predict disease diagnosis and provide a rapid diagnostic reference for doctors. In this study, we designed a BiGRU-Att-CapsNetwork model based on our proposed CMedBERT Chinese medical domain pre-trained language model to predict disease diagnosis in Chinese EMR. In the wide-ranging comparative experiments involving a real EMR dataset (SAHSU) and an academic evaluation task dataset (CCKS 2019), our model obtained competitive performance

    A Joint Model of Clinical Domain Classification and Slot Filling Based on RCNN and BiGRU-CRF

    Get PDF
    The task of the Intent Classification & Slot Filling serves as a key joint task in the voice assistant, which also plays the role of the pre-work in the construction of the medical consultation assistant system. How to distribute a doctor-patient conversation into a formatted electronic medical record to an accurate department (Intent Classification) to extract the key named entities or mentions (Slot Filling) through a specialized domain knowledge recognizer is one of the key steps of the entire system. In real cases, the medical vocabulary and clinical entities in different departments of the hospital often differ to some extent. Therefore, we propose a comprehensive model based on CMed-BERT, RCNN and BiGRU-CRF for a joint task of department identification and slot filling of the specific domain. Experimental results confirmed the competitiveness of our model
    corecore