123 research outputs found

    Exploring Partial Knowledge Base Inference in Biomedical Entity Linking

    Full text link
    Biomedical entity linking (EL) consists of named entity recognition (NER) and named entity disambiguation (NED). EL models are trained on corpora labeled by a predefined KB. However, it is a common scenario that only entities within a subset of the KB are precious to stakeholders. We name this scenario partial knowledge base inference: training an EL model with one KB and inferring on the part of it without further training. In this work, we give a detailed definition and evaluation procedures for this practically valuable but significantly understudied scenario and evaluate methods from three representative EL paradigms. We construct partial KB inference benchmarks and witness a catastrophic degradation in EL performance due to dramatically precision drop. Our findings reveal these EL paradigms can not correctly handle unlinkable mentions (NIL), so they are not robust to partial KB inference. We also propose two simple-and-effective redemption methods to combat the NIL issue with little computational overhead. Codes are released at https://github.com/Yuanhy1997/PartialKB-EL.Comment: Accepted by ACL-BioNLP 2023. The first two authors are contributed equall

    Large Language Models are Superpositions of All Characters: Attaining Arbitrary Role-play via Self-Alignment

    Full text link
    Considerable efforts have been invested in augmenting the role-playing proficiency of open-source large language models (LLMs) by emulating proprietary counterparts. Nevertheless, we posit that LLMs inherently harbor role-play capabilities, owing to the extensive knowledge of characters and potential dialogues ingrained in their vast training corpora. Thus, in this study, we introduce Ditto, a self-alignment method for role-play. Ditto capitalizes on character knowledge, encouraging an instruction-following LLM to simulate role-play dialogues as a variant of reading comprehension. This method creates a role-play training set comprising 4,000 characters, surpassing the scale of currently available datasets by tenfold regarding the number of roles. Subsequently, we fine-tune the LLM using this self-generated dataset to augment its role-playing capabilities. Upon evaluating our meticulously constructed and reproducible role-play benchmark and the roleplay subset of MT-Bench, Ditto, in various parameter scales, consistently maintains a consistent role identity and provides accurate role-specific knowledge in multi-turn role-play conversations. Notably, it outperforms all open-source role-play baselines, showcasing performance levels comparable to advanced proprietary chatbots. Furthermore, we present the first comprehensive cross-supervision alignment experiment in the role-play domain, revealing that the intrinsic capabilities of LLMs confine the knowledge within role-play. Meanwhile, the role-play styles can be easily acquired with the guidance of smaller models. We open-source related resources at https://github.com/OFA-Sys/Ditto

    Speculative Contrastive Decoding

    Full text link
    Large language models (LLMs) have shown extraordinary performance in various language tasks, but high computational requirements hinder their widespread deployment. Speculative decoding, which uses amateur models to predict the generation of expert models, has been proposed as a way to accelerate LLM inference. However, speculative decoding focuses on acceleration instead of making the best use of the token distribution from amateur models. We proposed Speculative Contrastive Decoding (SCD), an accelerated decoding method leveraging the natural contrast between expert and amateur models in speculative decoding. Comprehensive evaluations on four benchmarks show that SCD can achieve similar acceleration factors as speculative decoding while further improving the generation quality as the contrastive decoding. The analysis of token probabilities further demonstrates the compatibility between speculative and contrastive decoding. Overall, SCD provides an effective approach to enhance the decoding quality of LLMs while saving computational resources.Comment: Working in Progres

    Effect of directional solidification rate on the microstructure and properties of deformation-processed Cu-7Cr-0.1Ag in situ composites

    Get PDF
    The influence of directional solidification rate on the microstructure, mechanical properties and conductivity of deformation-processed Cu-7Cr-0.1Ag in situ composites produced by thermo-mechanical processing was systematically investigated. The microstructure was analyzed by optical microscopy and scanning electronic microscopy. The mechanical properties and conductivity were evaluated by tensile-testing machine and micro-ohmmeter, respectively. The results indicate that the size, shape and distribution of second-phase Cr grains are significantly different in the Cu-7Cr-0.1Ag alloys with different growth rates. At a growth rate of 200 μm s-1, the Cr grains transform into fine Cr fiber-like grains parallel to the pulling direction from the Cr dendrites. The tensile strength of the Cu-7Cr-0.1Ag in situ composites from the directional solidification (DS) alloys is significantly higher than that from the as-cast alloy, while the conductivity of the in situ composites from the DS alloys is slightly lower than that from the as-cast alloy. The following combinations of tensile strength, elongation to fracture and conductivity of the Cu-7Cr-0.1Ag in situ composites from the DS alloy with a growth rate of 200 μm s-1 and a cumulative cold deformation strain of 8 after isochronic aging treatment for 1 h can be obtained respectively as: (i) 1067 MPa, 2.9% and 74.9% IACS; or (ii) 1018 MPa, 3.0%, and 76.0% IACS or (iii) 906 MPa, 3.3% and 77.6% IACS

    Multi-hop Evidence Retrieval for Cross-document Relation Extraction

    Full text link
    Relation Extraction (RE) has been extended to cross-document scenarios because many relations are not simply described in a single document. This inevitably brings the challenge of efficient open-space evidence retrieval to support the inference of cross-document relations, along with the challenge of multi-hop reasoning on top of entities and evidence scattered in an open set of documents. To combat these challenges, we propose MR.COD (Multi-hop evidence retrieval for Cross-document relation extraction), which is a multi-hop evidence retrieval method based on evidence path mining and ranking. We explore multiple variants of retrievers to show evidence retrieval is essential in cross-document RE. We also propose a contextual dense retriever for this setting. Experiments on CodRED show that evidence retrieval with MR.COD effectively acquires crossdocument evidence and boosts end-to-end RE performance in both closed and open settings.Comment: ACL 2023 (Findings

    Self-Evolved Diverse Data Sampling for Efficient Instruction Tuning

    Full text link
    Enhancing the instruction-following ability of Large Language Models (LLMs) primarily demands substantial instruction-tuning datasets. However, the sheer volume of these imposes a considerable computational burden and annotation cost. To investigate a label-efficient instruction tuning method that allows the model itself to actively sample subsets that are equally or even more effective, we introduce a self-evolving mechanism DiverseEvol. In this process, a model iteratively augments its training subset to refine its own performance, without requiring any intervention from humans or more advanced LLMs. The key to our data sampling technique lies in the enhancement of diversity in the chosen subsets, as the model selects new data points most distinct from any existing ones according to its current embedding space. Extensive experiments across three datasets and benchmarks demonstrate the effectiveness of DiverseEvol. Our models, trained on less than 8% of the original dataset, maintain or improve performance compared with finetuning on full data. We also provide empirical evidence to analyze the importance of diversity in instruction data and the iterative scheme as opposed to one-time sampling. Our code is publicly available at https://github.com/OFA-Sys/DiverseEvol.git

    Experimental study on dynamic bending rigidity of aluminium cable steel reinforced

    Get PDF
    The dynamic bending rigidity of aluminium cable steel reinforced (ACSR) is a key parameter for analyzing the breeze vibration, galloping and de-icing vibration response of overhead transmission lines. In this paper, the calculation formula of the dynamic bending rigidity under axial force is derived based on the theory of Bernoulli-Euler beam. The dynamic response of one typical ACSR with different spans and different axial forces are studied by white-noise excitation and hammer excitation. The variation rules of dynamic bending rigidity of ACSR are presented. The comparison between the experimental results of the static and dynamic bending rigidity shows that the dynamic bending rigidity of the ACSR is much larger than the static bending rigidity

    PIVOINE: Instruction Tuning for Open-world Information Extraction

    Full text link
    We consider the problem of Open-world Information Extraction (Open-world IE), which extracts comprehensive entity profiles from unstructured texts. Different from the conventional closed-world setting of Information Extraction (IE), Open-world IE considers a more general situation where entities and relations could be beyond a predefined ontology. More importantly, we seek to develop a large language model (LLM) that is able to perform Open-world IE to extract desirable entity profiles characterized by (possibly fine-grained) natural language instructions. We achieve this by finetuning LLMs using instruction tuning. In particular, we construct INSTRUCTOPENWIKI, a substantial instruction tuning dataset for Open-world IE enriched with a comprehensive corpus, extensive annotations, and diverse instructions. We finetune the pretrained BLOOM models on INSTRUCTOPENWIKI and obtain PIVOINE, an LLM for Open-world IE with strong instruction-following capabilities. Our experiments demonstrate that PIVOINE significantly outperforms traditional closed-world methods and other LLM baselines, displaying impressive generalization capabilities on both unseen instructions and out-of-ontology cases. Consequently, PIVOINE emerges as a promising solution to tackle the open-world challenge in IE effectively
    • …
    corecore