56 research outputs found

    Towards Understanding the Capability of Large Language Models on Code Clone Detection: A Survey

    Full text link
    Code cloning, the duplication of code fragments, is common in software development. While some reuse aids productivity, excessive cloning hurts maintainability and introduces bugs. Hence, automatic code clone detection is vital. Meanwhile, large language models (LLMs) possess diverse code-related knowledge, making them versatile for various software engineering challenges. However, LLMs' performance in code clone detection is unclear and needs more study for accurate assessment. In this paper, we provide the first comprehensive evaluation of LLMs for clone detection, covering different clone types, languages, and prompts. We find advanced LLMs excel in detecting complex semantic clones, surpassing existing methods. Adding intermediate reasoning steps via chain-of-thought prompts noticeably enhances performance. Additionally, representing code as vector embeddings, especially with text encoders, effectively aids clone detection.Lastly, the ability of LLMs to detect code clones differs among various programming languages. Our study suggests that LLMs have potential for clone detection due to their language capabilities, offering insights for developing robust LLM-based methods to enhance software engineering.Comment: 13 pages, 3 figure

    Search for dark matter produced in association with bottom or top quarks in √s = 13 TeV pp collisions with the ATLAS detector

    Get PDF
    A search for weakly interacting massive particle dark matter produced in association with bottom or top quarks is presented. Final states containing third-generation quarks and miss- ing transverse momentum are considered. The analysis uses 36.1 fb−1 of proton–proton collision data recorded by the ATLAS experiment at √s = 13 TeV in 2015 and 2016. No significant excess of events above the estimated backgrounds is observed. The results are in- terpreted in the framework of simplified models of spin-0 dark-matter mediators. For colour- neutral spin-0 mediators produced in association with top quarks and decaying into a pair of dark-matter particles, mediator masses below 50 GeV are excluded assuming a dark-matter candidate mass of 1 GeV and unitary couplings. For scalar and pseudoscalar mediators produced in association with bottom quarks, the search sets limits on the production cross- section of 300 times the predicted rate for mediators with masses between 10 and 50 GeV and assuming a dark-matter mass of 1 GeV and unitary coupling. Constraints on colour- charged scalar simplified models are also presented. Assuming a dark-matter particle mass of 35 GeV, mediator particles with mass below 1.1 TeV are excluded for couplings yielding a dark-matter relic density consistent with measurements

    Statistical Inferences and Applications of the Half Exponential Power Distribution

    No full text
    We investigate the statistical inferences and applications of the half exponential power distribution for the first time. The proposed model defined on the nonnegative reals extends the half normal distribution and is more flexible. The characterizations and properties involving moments and some measures based on moments of this distribution are derived. The inference aspects using methods of moment and maximum likelihood are presented. We also study the performance of the estimators using the Monte Carlo simulation. Finally, we illustrate it with two real applications

    Estimation and Prediction for Gompertz Distribution under General Progressive Censoring

    No full text
    In this article, we discuss the estimation of the parameters for Gompertz distribution and prediction using general progressive Type-II censoring. Based on the Expectation–Maximization algorithm, we calculate the maximum likelihood estimates. Bayesian estimates are considered under different loss functions, which are symmetrical, asymmetrical and balanced, respectively. An approximate method—Tierney and Kadane—is used to derive the estimates. Besides, the Metropolis-Hasting (MH) algorithm is applied to get the Bayesian estimates as well. According to Fisher information matrix, we acquire asymptotic confidence intervals. Bootstrap intervals are also established. Furthermore, we build the highest posterior density intervals through the sample generated by the MH algorithm. Then, Bayesian predictive intervals and estimates for future samples are provided. Finally, for evaluating the quality of the approaches, a numerical simulation study is implemented. In addition, we analyze two real datasets

    Bayesian Estimation of Entropy for Burr Type XII Distribution under Progressive Type-II Censored Data

    No full text
    With the rapid development of statistics, information entropy is proposed as an important indicator used to quantify information uncertainty. In this paper, maximum likelihood and Bayesian methods are used to obtain the estimators of the entropy for a two-parameter Burr type XII distribution under progressive type-II censored data. In the part of maximum likelihood estimation, the asymptotic confidence intervals of entropy are calculated. In Bayesian estimation, we consider non-informative and informative priors respectively, and asymmetric and symmetric loss functions are both adopted. Meanwhile, the posterior risk is also calculated to evaluate the performances of the entropy estimators against different loss functions. In a numerical simulation, the Lindley approximation and the Markov chain Monte Carlo method were used to obtain the Bayesian estimates. In turn, the highest posterior density credible intervals of the entropy were derived. Finally, average absolute bias and mean square error were used to evaluate the estimators under different methods, and a real dataset was selected to illustrate the feasibility of the above estimation model

    Statistical Inference of the Lifetime Performance Index with the Log-Logistic Distribution Based on Progressive First-Failure-Censored Data

    No full text
    Estimating the accurate evaluation of product lifetime performance has always been a hot topic in manufacturing industry. This paper, based on the lifetime performance index, focuses on its evaluation when a lower specification limit is given. The progressive first-failure-censored data we discuss have a common log-logistic distribution. Both Bayesian and non-Bayesian method are studied. Bayes estimator of the parameters of the log-logistic distribution and the lifetime performance index are obtained using both the Lindley approximation and Monte Carlo Markov Chain methods under symmetric and asymmetric loss functions. As for interval estimation, we apply the maximum likelihood estimator to construct the asymptotic confidence intervals and the Metropolis–Hastings algorithm to establish the highest posterior density credible intervals. Moreover, we analyze a real data set for demonstrative purposes. In addition, different criteria for deciding the optimal censoring scheme have been studied

    Statistical Inference of Weighted Exponential Distribution under Joint Progressive Type-II Censoring

    No full text
    The weighted exponential distribution is a promising skewed distribution in the life-testing experiment. The joint progressive type-II censoring (JPC) scheme is an effective approach to reducing costs. In this paper, we consider the estimates of parameters of the weighted exponential distribution with the JPC data. Two populations, whose scale parameters are the same but the shape parameters of which are different, are chosen to be studied. We first evaluate the parameters with the maximum likelihood method. Because the maximum likelihood estimates of parameters cannot be obtained in closed form, we apply the Newton–Raphson method in this part. Bayesian estimates and the corresponding credible intervals under the squared error loss function are computed by using the Markov Chain Monte Carlo method. After that, we use the bootstrap method to calculate the associated confidence intervals of the unknown parameters. A simulation has been performed to test the feasibility of the above methods and real data analysis is also provided for illustrative purposes

    Statistical Inference and Optimal Design of Accelerated Life Testing for the Chen Distribution under Progressive Type-II Censoring

    No full text
    This paper discusses statistical inference and optimal design of constant-stress accelerated life testing for the Chen distribution under progressive Type-II censoring. The scale parameter of the life distribution is assumed to be a logarithmic linear function of the stress level. The maximum likelihood estimates of the parameters are obtained. Then, the observed Fisher information matrix is derived and utilized to construct asymptotic confidence intervals. Meanwhile, the parametric bootstrap methods are provided for the interval estimation. In addition, the Bayes estimates under the squared error loss function are obtained by applying the Tierney and Kadane technique and Lindley’s approximation. As for the optimal design, D- and A-optimality criteria are considered to determine the optimal transformed stress level. Finally, the simulation is carried out to demonstrate the proposed estimation techniques and the optimal criteria, and a real data set is discussed
    corecore