238 research outputs found

    Efficient and Effective Text Encoding for Chinese LLaMA and Alpaca

    Full text link
    Large Language Models (LLMs), such as ChatGPT and GPT-4, have dramatically transformed natural language processing research and shown promising strides towards Artificial General Intelligence (AGI). Nonetheless, the high costs associated with training and deploying LLMs present substantial obstacles to transparent, accessible academic research. While several large language models, such as LLaMA, have been open-sourced by the community, these predominantly focus on English corpora, limiting their usefulness for other languages. In this paper, we propose a method to augment LLaMA with capabilities for understanding and generating Chinese text and its ability to follow instructions. We achieve this by extending LLaMA's existing vocabulary with an additional 20,000 Chinese tokens, thereby improving its encoding efficiency and semantic understanding of Chinese. We further incorporate secondary pre-training using Chinese data and fine-tune the model with Chinese instruction datasets, significantly enhancing the model's ability to comprehend and execute instructions. Our experimental results indicate that the newly proposed model markedly enhances the original LLaMA's proficiency in understanding and generating Chinese content. Additionally, the results on the C-Eval dataset yield competitive performance among the models with several times the size of ours. We have made our pre-trained models, training scripts, and other resources available through GitHub, fostering open research for our community. Chinese LLaMA series: \url{https://github.com/ymcui/Chinese-LLaMA-Alpaca} and Chinese Llama-2 series: \url{https://github.com/ymcui/Chinese-LLaMA-Alpaca-2}Comment: 21 page

    IDOL: Indicator-oriented Logic Pre-training for Logical Reasoning

    Full text link
    In the field of machine reading comprehension (MRC), existing systems have surpassed the average performance of human beings in many tasks like SQuAD. However, there is still a long way to go when it comes to logical reasoning. Although some methods for it have been put forward, they either are designed in a quite complicated way or rely too much on external structures. In this paper, we proposed IDOL (InDicator-Oriented Logic Pre-training), an easy-to-understand but highly effective further pre-training task which logically strengthens the pre-trained models with the help of 6 types of logical indicators and a logically rich dataset LGP (LoGic Pre-training). IDOL achieves state-of-the-art performance on ReClor and LogiQA, the two most representative benchmarks in logical reasoning MRC, and is proven to be capable of generalizing to different pre-trained models and other types of MRC benchmarks like RACE and SQuAD 2.0 while keeping competitive general language understanding ability through testing on tasks in GLUE. Besides, at the beginning of the era of large language models, we take several of them like ChatGPT into comparison and find that IDOL still shows its advantage.Comment: Accepted to the Findings of ACL 202

    Preliminary Study on Functional and Aesthetic Reconstruction by Using a Small Artery-only Free Medial Flap of the Second Toe for Fingertip Injuries

    Get PDF
    OBJECTIVES: This study was designed to introduce the feasibility of fingertip reconstruction by using a free medial flap of the second toe without vein anastomosis. METHODS: In total, 8 patients with fingertip injuries were treated successfully with this method. Patients who underwent reconstruction from September 2016 to October 2017 in our hospital with an artery-only free medial flap transfer of the second toe for fingertip injuries were included, and patients who underwent additional procedures that may impact the postoperative results and were followed up for less than 6 months were excluded. Clinical trial registration: ChiCTR19000021883. RESULTS: According to the Allen classification, five patients had Type 3 injuries, and three patients had Type 4 injuries. One arterial nerve and one digital nerve were repaired at the same time. No additional dissection was performed in either the donor or recipient site of the dorsal or volar vein. Postoperative venous congestion was monitored based on the color, temperature and the degree of tissue oxygen saturation. The flap size ranged from 1.20*1.0 cm2 to 1.80*1.0 cm2 . The reconstruction time was 71.86 (SD 14.75) minutes. The two-point discrimination and the monofilament results were satisfying; cold intolerance did not appear in five patients, and the other three patients had cold intolerance with grades of 4, 12 and 26, which were considered satisfactory. Moreover, leech therapy, continuous bleeding and needle sutures were not utilized in any cases. CONCLUSIONS: Reconstruction with a small artery-only free medial flap transfer of the second toe led to satisfactory sensory and motor function in the selected patients with fingertip injuries

    Unsupervised Behavior Extraction via Random Intent Priors

    Full text link
    Reward-free data is abundant and contains rich prior knowledge of human behaviors, but it is not well exploited by offline reinforcement learning (RL) algorithms. In this paper, we propose UBER, an unsupervised approach to extract useful behaviors from offline reward-free datasets via diversified rewards. UBER assigns different pseudo-rewards sampled from a given prior distribution to different agents to extract a diverse set of behaviors, and reuse them as candidate policies to facilitate the learning of new tasks. Perhaps surprisingly, we show that rewards generated from random neural networks are sufficient to extract diverse and useful behaviors, some even close to expert ones. We provide both empirical and theoretical evidence to justify the use of random priors for the reward function. Experiments on multiple benchmarks showcase UBER's ability to learn effective and diverse behavior sets that enhance sample efficiency for online RL, outperforming existing baselines. By reducing reliance on human supervision, UBER broadens the applicability of RL to real-world scenarios with abundant reward-free data.Comment: Thirty-seventh Conference on Neural Information Processing System

    Data Poisoning Attacks Against Multimodal Encoders

    Get PDF
    Traditional machine learning (ML) models usually rely on large-scale labeled datasets to achieve strong performance. However, such labeled datasets are often challenging and expensive to obtain. Also, the predefined categories limit the model's ability to generalize to other visual concepts as additional labeled data is required. On the contrary, the newly emerged multimodal model, which contains both visual and linguistic modalities, learns the concept of images from the raw text. It is a promising way to solve the above problems as it can use easy-to-collect image-text pairs to construct the training dataset and the raw texts contain almost unlimited categories according to their semantics. However, learning from a large-scale unlabeled dataset also exposes the model to the risk of potential poisoning attacks, whereby the adversary aims to perturb the model's training dataset to trigger malicious behaviors in it. Previous work mainly focuses on the visual modality. In this paper, we instead focus on answering two questions: (1) Is the linguistic modality also vulnerable to poisoning attacks? and (2) Which modality is most vulnerable? To answer the two questions, we conduct three types of poisoning attacks against CLIP, the most representative multimodal contrastive learning framework. Extensive evaluations on different datasets and model architectures show that all three attacks can perform well on the linguistic modality with only a relatively low poisoning rate and limited epochs. Also, we observe that the poisoning effect differs between different modalities, i.e., with lower MinRank in the visual modality and with higher Hit@K when K is small in the linguistic modality. To mitigate the attacks, we propose both pre-training and post-training defenses. We empirically show that both defenses can significantly reduce the attack performance while preserving the model's utility
    corecore