320 research outputs found

    What Makes Good In-context Demonstrations for Code Intelligence Tasks with LLMs?

    Full text link
    Pre-trained models of source code have gained widespread popularity in many code intelligence tasks. Recently, with the scaling of the model and corpus size, large language models have shown the ability of in-context learning (ICL). ICL employs task instructions and a few examples as demonstrations, and then inputs the demonstrations to the language models for making predictions. This new learning paradigm is training-free and has shown impressive performance in various natural language processing and code intelligence tasks. However, the performance of ICL heavily relies on the quality of demonstrations, e.g., the selected examples. It is important to systematically investigate how to construct a good demonstration for code-related tasks. In this paper, we empirically explore the impact of three key factors on the performance of ICL in code intelligence tasks: the selection, order, and number of demonstration examples. We conduct extensive experiments on three code intelligence tasks including code summarization, bug fixing, and program synthesis. Our experimental results demonstrate that all the above three factors dramatically impact the performance of ICL in code intelligence tasks. Additionally, we summarize our findings and provide takeaway suggestions on how to construct effective demonstrations, taking into account these three perspectives. We also show that a carefully-designed demonstration based on our findings can lead to substantial improvements over widely-used demonstration construction methods, e.g., improving BLEU-4, EM, and EM by at least 9.90%, 175.96%, and 50.81% on code summarization, bug fixing, and program synthesis, respectivelyComment: This paper is accepted by ASE 202

    Scaling Law of Large Sequential Recommendation Models

    Full text link
    Scaling of neural networks has recently shown great potential to improve the model capacity in various fields. Specifically, model performance has a power-law relationship with model size or data size, which provides important guidance for the development of large-scale models. However, there is still limited understanding on the scaling effect of user behavior models in recommender systems, where the unique data characteristics (e.g. data scarcity and sparsity) pose new challenges to explore the scaling effect in recommendation tasks. In this work, we focus on investigating the scaling laws in large sequential recommendation models. Specially, we consider a pure ID-based task formulation, where the interaction history of a user is formatted as a chronological sequence of item IDs. We don't incorporate any side information (e.g. item text), because we would like to explore how scaling law holds from the perspective of user behavior. With specially improved strategies, we scale up the model size to 0.8B parameters, making it feasible to explore the scaling effect in a diverse range of model sizes. As the major findings, we empirically show that scaling law still holds for these trained models, even in data-constrained scenarios. We then fit the curve for scaling law, and successfully predict the test loss of the two largest tested model scales. Furthermore, we examine the performance advantage of scaling effect on five challenging recommendation tasks, considering the unique issues (e.g. cold start, robustness, long-term preference) in recommender systems. We find that scaling up the model size can greatly boost the performance on these challenging tasks, which again verifies the benefits of large recommendation models

    Segmentation of kidney lesions with attention model based on Deeplab

    Get PDF
    We participate this challenge by developing a hierarchical framework. We build the model from two fully convolutional networks: (1) a simple Unet model to normalize the input iamges, (2) a segmentaion network which is an attention model based on Deeplab model. Two models are connected in tandem and trained end-to-end. To ensure a better results, we use the preprocess method proposed by nnUnet in our experiments

    Toll-like receptor 9 negatively regulates pancreatic islet beta cell growth and function in a mouse model of type 1 diabetes

    Get PDF
    Aims/hypothesis Innate immune effectors interact with the environment to contribute to the pathogenesis of the autoimmune disease, type 1 diabetes. Although recent studies have suggested that innate immune Toll-like receptors (TLRs) are involved in tissue development, little is known about the role of TLRs in tissue development, compared with autoimmunity. We aimed to fill the knowledge gap by investigating the role of TLR9 in the development and function of islet beta cells in type 1 diabetes, using NOD mice. Methods We generated Tlr9−/− NOD mice and examined them for type 1 diabetes development and beta cell function, including insulin secretion and glucose tolerance. We assessed islet and beta cell number and characterised CD140a expression on beta cells by flow cytometry. We also tested beta cell function in Tlr9−/− C57BL/6 mice. Finally, we used TLR9 antagonists to block TLR9 signalling in wild-type NOD mice to verify the role of TLR9 in beta cell development and function. Results TLR9 deficiency promoted pancreatic islet development and beta cell differentiation, leading to enhanced glucose tolerance, improved insulin sensitivity and enhanced first-phase insulin secretory response. This was, in part, mediated by upregulation of CD140a (also known as platelet-derived growth factor receptor-α [PDGFRα]). In the absence of TLR9, induced by either genetic targeting or treatment with TLR9 antagonists, which had similar effects on ontogenesis and function of beta cells, NOD mice were protected from diabetes. Conclusions/interpretation Our study links TLR9 and the CD140a pathway in regulating islet beta cell development and function and indicates a potential therapeutic target for diabetes prevention and/or treatment
    • …
    corecore