238 research outputs found
Efficient and Effective Text Encoding for Chinese LLaMA and Alpaca
Large Language Models (LLMs), such as ChatGPT and GPT-4, have dramatically
transformed natural language processing research and shown promising strides
towards Artificial General Intelligence (AGI). Nonetheless, the high costs
associated with training and deploying LLMs present substantial obstacles to
transparent, accessible academic research. While several large language models,
such as LLaMA, have been open-sourced by the community, these predominantly
focus on English corpora, limiting their usefulness for other languages. In
this paper, we propose a method to augment LLaMA with capabilities for
understanding and generating Chinese text and its ability to follow
instructions. We achieve this by extending LLaMA's existing vocabulary with an
additional 20,000 Chinese tokens, thereby improving its encoding efficiency and
semantic understanding of Chinese. We further incorporate secondary
pre-training using Chinese data and fine-tune the model with Chinese
instruction datasets, significantly enhancing the model's ability to comprehend
and execute instructions. Our experimental results indicate that the newly
proposed model markedly enhances the original LLaMA's proficiency in
understanding and generating Chinese content. Additionally, the results on the
C-Eval dataset yield competitive performance among the models with several
times the size of ours. We have made our pre-trained models, training scripts,
and other resources available through GitHub, fostering open research for our
community. Chinese LLaMA series:
\url{https://github.com/ymcui/Chinese-LLaMA-Alpaca} and Chinese Llama-2 series:
\url{https://github.com/ymcui/Chinese-LLaMA-Alpaca-2}Comment: 21 page
IDOL: Indicator-oriented Logic Pre-training for Logical Reasoning
In the field of machine reading comprehension (MRC), existing systems have
surpassed the average performance of human beings in many tasks like SQuAD.
However, there is still a long way to go when it comes to logical reasoning.
Although some methods for it have been put forward, they either are designed in
a quite complicated way or rely too much on external structures. In this paper,
we proposed IDOL (InDicator-Oriented Logic Pre-training), an easy-to-understand
but highly effective further pre-training task which logically strengthens the
pre-trained models with the help of 6 types of logical indicators and a
logically rich dataset LGP (LoGic Pre-training). IDOL achieves state-of-the-art
performance on ReClor and LogiQA, the two most representative benchmarks in
logical reasoning MRC, and is proven to be capable of generalizing to different
pre-trained models and other types of MRC benchmarks like RACE and SQuAD 2.0
while keeping competitive general language understanding ability through
testing on tasks in GLUE. Besides, at the beginning of the era of large
language models, we take several of them like ChatGPT into comparison and find
that IDOL still shows its advantage.Comment: Accepted to the Findings of ACL 202
Preliminary Study on Functional and Aesthetic Reconstruction by Using a Small Artery-only Free Medial Flap of the Second Toe for Fingertip Injuries
OBJECTIVES: This study was designed to introduce the feasibility of fingertip reconstruction by using a free medial flap of the second toe without vein anastomosis. METHODS: In total, 8 patients with fingertip injuries were treated successfully with this method. Patients who underwent reconstruction from September 2016 to October 2017 in our hospital with an artery-only free medial flap transfer of the second toe for fingertip injuries were included, and patients who underwent additional procedures that may impact the postoperative results and were followed up for less than 6 months were excluded. Clinical trial registration: ChiCTR19000021883. RESULTS: According to the Allen classification, five patients had Type 3 injuries, and three patients had Type 4 injuries. One arterial nerve and one digital nerve were repaired at the same time. No additional dissection was performed in either the donor or recipient site of the dorsal or volar vein. Postoperative venous congestion was monitored based on the color, temperature and the degree of tissue oxygen saturation. The flap size ranged from 1.20*1.0 cm2 to 1.80*1.0 cm2 . The reconstruction time was 71.86 (SD 14.75) minutes. The two-point discrimination and the monofilament results were satisfying; cold intolerance did not appear in five patients, and the other three patients had cold intolerance with grades of 4, 12 and 26, which were considered satisfactory. Moreover, leech therapy, continuous bleeding and needle sutures were not utilized in any cases. CONCLUSIONS: Reconstruction with a small artery-only free medial flap transfer of the second toe led to satisfactory sensory and motor function in the selected patients with fingertip injuries
Unsupervised Behavior Extraction via Random Intent Priors
Reward-free data is abundant and contains rich prior knowledge of human
behaviors, but it is not well exploited by offline reinforcement learning (RL)
algorithms. In this paper, we propose UBER, an unsupervised approach to extract
useful behaviors from offline reward-free datasets via diversified rewards.
UBER assigns different pseudo-rewards sampled from a given prior distribution
to different agents to extract a diverse set of behaviors, and reuse them as
candidate policies to facilitate the learning of new tasks. Perhaps
surprisingly, we show that rewards generated from random neural networks are
sufficient to extract diverse and useful behaviors, some even close to expert
ones. We provide both empirical and theoretical evidence to justify the use of
random priors for the reward function. Experiments on multiple benchmarks
showcase UBER's ability to learn effective and diverse behavior sets that
enhance sample efficiency for online RL, outperforming existing baselines. By
reducing reliance on human supervision, UBER broadens the applicability of RL
to real-world scenarios with abundant reward-free data.Comment: Thirty-seventh Conference on Neural Information Processing System
Data Poisoning Attacks Against Multimodal Encoders
Traditional machine learning (ML) models usually rely on large-scale labeled
datasets to achieve strong performance. However, such labeled datasets are
often challenging and expensive to obtain. Also, the predefined categories
limit the model's ability to generalize to other visual concepts as additional
labeled data is required. On the contrary, the newly emerged multimodal model,
which contains both visual and linguistic modalities, learns the concept of
images from the raw text. It is a promising way to solve the above problems as
it can use easy-to-collect image-text pairs to construct the training dataset
and the raw texts contain almost unlimited categories according to their
semantics. However, learning from a large-scale unlabeled dataset also exposes
the model to the risk of potential poisoning attacks, whereby the adversary
aims to perturb the model's training dataset to trigger malicious behaviors in
it. Previous work mainly focuses on the visual modality. In this paper, we
instead focus on answering two questions: (1) Is the linguistic modality also
vulnerable to poisoning attacks? and (2) Which modality is most vulnerable? To
answer the two questions, we conduct three types of poisoning attacks against
CLIP, the most representative multimodal contrastive learning framework.
Extensive evaluations on different datasets and model architectures show that
all three attacks can perform well on the linguistic modality with only a
relatively low poisoning rate and limited epochs. Also, we observe that the
poisoning effect differs between different modalities, i.e., with lower MinRank
in the visual modality and with higher Hit@K when K is small in the linguistic
modality. To mitigate the attacks, we propose both pre-training and
post-training defenses. We empirically show that both defenses can
significantly reduce the attack performance while preserving the model's
utility
- …