4,764 research outputs found
MeDAL: Medical Abbreviation Disambiguation Dataset for Natural Language Understanding Pretraining
One of the biggest challenges that prohibit the use of many current NLP
methods in clinical settings is the availability of public datasets. In this
work, we present MeDAL, a large medical text dataset curated for abbreviation
disambiguation, designed for natural language understanding pre-training in the
medical domain. We pre-trained several models of common architectures on this
dataset and empirically showed that such pre-training leads to improved
performance and convergence speed when fine-tuning on downstream medical tasks.Comment: EMNLP 2020 Clinical NL
The StatCan Dialogue Dataset: Retrieving Data Tables through Conversations with Genuine Intents
We introduce the StatCan Dialogue Dataset consisting of 19,379 conversation
turns between agents working at Statistics Canada and online users looking for
published data tables. The conversations stem from genuine intents, are held in
English or French, and lead to agents retrieving one of over 5000 complex data
tables. Based on this dataset, we propose two tasks: (1) automatic retrieval of
relevant tables based on a on-going conversation, and (2) automatic generation
of appropriate agent responses at each turn. We investigate the difficulty of
each task by establishing strong baselines. Our experiments on a temporal data
split reveal that all models struggle to generalize to future conversations, as
we observe a significant drop in performance across both tasks when we move
from the validation to the test set. In addition, we find that response
generation models struggle to decide when to return a table. Considering that
the tasks pose significant challenges to existing models, we encourage the
community to develop models for our task, which can be directly used to help
knowledge workers find relevant tables for live chat users.Comment: Accepted at EACL 202
GaAs-InGaAs-GaAs fin-array tunnel diodes on (001) Si substrates with room-temperature peak-to-valley current ratio of 5.4
In this letter, we report the selective area growth of GaAs, In0.2Ga0.8As, and GaAs/In0.2Ga0.8As/GaAs quantum-well fins of 65-nm width on exactly orientated (001) Si substrates. By exploiting high aspect ratio trenches formed by patterned SiO2 on Si and a V-grooved Si (111) surface in the aspect ratio trapping process, we are able to achieve good material quality and structural properties, as evidenced by x-ray diffraction, scanning electron microscopy, and transmission electron microscopy. The fabricated GaAs-In0.2Ga0.8As-GaAs fin-array tunnel diodes exhibit a maximum room-temperature peak-to-valley current ratio of 5.4, and negative differential resistance characteristics up to 200 °C
Evaluating Correctness and Faithfulness of Instruction-Following Models for Question Answering
Retriever-augmented instruction-following models are attractive alternatives
to fine-tuned approaches for information-seeking tasks such as question
answering (QA). By simply prepending retrieved documents in its input along
with an instruction, these models can be adapted to various information domains
and tasks without additional fine-tuning. While the model responses tend to be
natural and fluent, the additional verbosity makes traditional QA evaluation
metrics such as exact match (EM) and F1 unreliable for accurately quantifying
model performance.
In this work, we investigate the performance of instruction-following models
across three information-seeking QA tasks. We use both automatic and human
evaluation to evaluate these models along two dimensions: 1) how well they
satisfy the user's information need (correctness), and 2) whether they produce
a response based on the provided knowledge (faithfulness). Guided by human
evaluation and analysis, we highlight the shortcomings of traditional metrics
for both correctness and faithfulness. We then propose simple token-overlap
based and model-based metrics that reflect the true performance of these
models. Our analysis reveals that instruction-following models are competitive,
and sometimes even outperform fine-tuned models for correctness. However, these
models struggle to stick to the provided knowledge and often hallucinate in
their responses. We hope our work encourages a more holistic evaluation of
instruction-following models for QA. Our code and data is available at
https://github.com/McGill-NLP/instruct-q
FIXED: Frustratingly Easy Domain Generalization with Mixup
Domain generalization (DG) aims to learn a generalizable model from multiple
training domains such that it can perform well on unseen target domains. A
popular strategy is to augment training data to benefit generalization through
methods such as Mixup~\cite{zhang2018mixup}. While the vanilla Mixup can be
directly applied, theoretical and empirical investigations uncover several
shortcomings that limit its performance. Firstly, Mixup cannot effectively
identify the domain and class information that can be used for learning
invariant representations. Secondly, Mixup may introduce synthetic noisy data
points via random interpolation, which lowers its discrimination capability.
Based on the analysis, we propose a simple yet effective enhancement for
Mixup-based DG, namely domain-invariant Feature mIXup (FIX). It learns
domain-invariant representations for Mixup. To further enhance discrimination,
we leverage existing techniques to enlarge margins among classes to further
propose the domain-invariant Feature MIXup with Enhanced Discrimination (FIXED)
approach. We present theoretical insights about guarantees on its
effectiveness. Extensive experiments on seven public datasets across two
modalities including image classification (Digits-DG, PACS, Office-Home) and
time series (DSADS, PAMAP2, UCI-HAR, and USC-HAD) demonstrate that our approach
significantly outperforms nine state-of-the-art related methods, beating the
best performing baseline by 6.5\% on average in terms of test accuracy. Code is
available at:
https://github.com/jindongwang/transferlearning/tree/master/code/deep/fixed.Comment: First Conference on Parsimony and Learning (CPAL) 2024; code for DG
at: https://github.com/jindongwang/transferlearning/tree/master/code/DeepD
Generation of Oligodendrocyte Progenitor Cells From Mouse Bone Marrow Cells.
Oligodendrocyte progenitor cells (OPCs) are a subtype of glial cells responsible for myelin regeneration. Oligodendrocytes (OLGs) originate from OPCs and are the myelinating cells in the central nervous system (CNS). OLGs play an important role in the context of lesions in which myelin loss occurs. Even though many protocols for isolating OPCs have been published, their cellular yield remains a limit for clinical application. The protocol proposed here is novel and has practical value; in fact, OPCs can be generated from a source of autologous cells without gene manipulation. Our method represents a rapid, and high-efficiency differentiation protocol for generating mouse OLGs from bone marrow-derived cells using growth-factor defined media. With this protocol, it is possible to obtain mature OLGs in 7-8 weeks. Within 2-3 weeks from bone marrow (BM) isolation, after neurospheres formed, the cells differentiate into Nestin+ Sox2+ neural stem cells (NSCs), around 30 days. OPCs specific markers start to be expressed around day 38, followed by RIP+O4+ around day 42. CNPase+ mature OLGs are finally obtained around 7-8 weeks. Further, bone marrow-derived OPCs exhibited therapeutic effect in shiverer (Shi) mice, promoting myelin regeneration and reducing the tremor. Here, we propose a method by which OLGs can be generated starting from BM cells and have similar abilities to subventricular zone (SVZ)-derived cells. This protocol significantly decreases the timing and costs of the OLGs differentiation within 2 months of culture
Identification of wheat seedling varieties based on MssiapNet
IntroductionIn the actual planting of wheat, there are often shortages of seedlings and broken seedlings on long ridges in the field, thus affecting grain yield and indirectly causing economic losses. Variety identification of wheat seedlings using physical methods timeliness and is unsuitable for universal dissemination. Recognition of wheat seedling varieties using deep learning models has high timeliness and accuracy, but fewer researchers exist. Therefore, in this paper, a lightweight wheat seedling variety recognition model, MssiapNet, is proposed.MethodsThe model is based on the MobileVit-XS and increases the model's sensitivity to subtle differences between different varieties by introducing the scSE attention mechanism in the MV2 module, so the recognition accuracy is improved. In addition, this paper proposes the IAP module to fuse the identified feature information. Subsequently, training was performed on a self-constructed real dataset, which included 29,020 photographs of wheat seedlings of 29 varieties.ResultsThe recognition accuracy of this model is 96.85%, which is higher than the other nine mainstream classification models. Although it is only 0.06 higher than the Resnet34 model, the number of parameters is only 1/3 of that. The number of parameters required for MssiapNet is 29.70MB, and the single image Execution time and the single image Delay time are 0.16s and 0.05s. The MssiapNet was visualized, and the heat map showed that the model was superior for wheat seedling variety identification compared with MobileVit-XS.DiscussionThe proposed model has a good recognition effect on wheat seedling varieties and uses a few parameters with fast inference speed, which makes it easy to be subsequently deployed on mobile terminals for practical performance testing
- …