220 research outputs found
Learning Cross-modal Context Graph for Visual Grounding
Visual grounding is a ubiquitous building block in many vision-language tasks
and yet remains challenging due to large variations in visual and linguistic
features of grounding entities, strong context effect and the resulting
semantic ambiguities. Prior works typically focus on learning representations
of individual phrases with limited context information. To address their
limitations, this paper proposes a language-guided graph representation to
capture the global context of grounding entities and their relations, and
develop a cross-modal graph matching strategy for the multiple-phrase visual
grounding task. In particular, we introduce a modular graph neural network to
compute context-aware representations of phrases and object proposals
respectively via message propagation, followed by a graph-based matching module
to generate globally consistent localization of grounding phrases. We train the
entire graph neural network jointly in a two-stage strategy and evaluate it on
the Flickr30K Entities benchmark. Extensive experiments show that our method
outperforms the prior state of the arts by a sizable margin, evidencing the
efficacy of our grounding framework. Code is available at
"https://github.com/youngfly11/LCMCG-PyTorch".Comment: AAAI-202
Automatic Generation of Electronic Medical Record Based on GPT2 Model
Writing Electronic Medical Records (EMR) as one of daily major tasks of doctors, consumes a lot of time and effort from doctors. This paper reports our efforts to generate electronic medical records using the language model. Through the training of massive real-world EMR data, the CMedGPT2 model provided by us can achieve the ideal Chinese electronic medical record generation. The experimental results prove that the generated electronic medical record text can be applied to the auxiliary medical record work to reduce the burden on the compose and provide a fast and accurate reference for composing work
An Word2vec based on Chinese Medical Knowledge
Introducing a large amount of external prior domain knowledge will effectively improve the performance of the word embedded language model in downstream NLP tasks. Based on this assumption, we collect and collate a medical corpus data with about 36M (Million) characters and use the data of CCKS2019 as the test set to carry out multiple classifications and named entity recognition (NER) tasks with the generated word and character vectors. Compared with the results of BERT, our models obtained the ideal performance and efficiency results
Disease Diagnosis Prediction of EMR Based on BiGRL-Att-CapsNetwork Model
Electronic Medical Records (EMR) carry a large number of diseases characteristics, history and other specific details of patients, which has great value for medical diagnosis. These data with diagnostic labels can help automated diagnostic assistant to predict disease diagnosis and provide a rapid diagnostic reference for doctors. In this study, we designed a BiGRU-Att-CapsNetwork model based on our proposed CMedBERT Chinese medical domain pre-trained language model to predict disease diagnosis in Chinese EMR. In the wide-ranging comparative experiments involving a real EMR dataset (SAHSU) and an academic evaluation task dataset (CCKS 2019), our model obtained competitive performance
A Joint Model of Clinical Domain Classification and Slot Filling Based on RCNN and BiGRU-CRF
The task of the Intent Classification & Slot Filling serves as a key joint task in the voice assistant, which also plays the role of the pre-work in the construction of the medical consultation assistant system. How to distribute a doctor-patient conversation into a formatted electronic medical record to an accurate department (Intent Classification) to extract the key named entities or mentions (Slot Filling) through a specialized domain knowledge recognizer is one of the key steps of the entire system. In real cases, the medical vocabulary and clinical entities in different departments of the hospital often differ to some extent. Therefore, we propose a comprehensive model based on CMed-BERT, RCNN and BiGRU-CRF for a joint task of department identification and slot filling of the specific domain. Experimental results confirmed the competitiveness of our model
Recommended from our members
The loss of taste genes in cetaceans
Background: Five basic taste modalities, sour, sweet, bitter, salt and umami, can be distinguished by humans and are fundamental for physical and ecological adaptations in mammals. Molecular genetic studies of the receptor genes for these tastes have been conducted in terrestrial mammals; however, little is known about the evolution and adaptation of these genes in marine mammals. Results: Here, all five basic taste modalities, sour, sweet, bitter, salt and umami, were investigated in cetaceans. The sequence characteristics and evolutionary analyses of taste receptor genes suggested that nearly all cetaceans may have lost all taste modalities except for that of salt. Conclusions: This is the first study to comprehensively examine the five basic taste modalities in cetaceans with extensive taxa sampling. Our results suggest that cetaceans have lost four of the basic taste modalities including sour, sweet, umami, and most of the ability to sense bitter tastes. The integrity of the candidate salt taste receptor genes in all the cetaceans examined may be because of their function in Na+ reabsorption, which is key to osmoregulation and aquatic adaptation. Electronic supplementary material The online version of this article (doi:10.1186/s12862-014-0218-8) contains supplementary material, which is available to authorized users
- …