20 research outputs found

    An Experimental Study on Attribute Validity of Code Quality Evaluation Model

    Get PDF
    Regarding the practicality of the quality evaluation model, the lack of quantitative experimental evaluation affects the effective use of the quality model, and also a lack of effective guidance for choosing the model. Aiming at this problem, based on the sensitivity of the quality evaluation model to code defects, a machine learning-based quality evaluation attribute validity verification method is proposed. This method conducts comparative experiments by controlling variables. First, extract the basic metric elements; then, convert them into quality attributes of the software; finally, to verify the quality evaluation model and the effectiveness of medium quality attributes, this paper compares machine learning methods based on quality attributes with those based on text features, and conducts experimental evaluation in two data sets. The result shows that the effectiveness of quality attributes under control variables is better, and leads by 15% in AdaBoostClassifier; when the text feature extraction method is increased to 50 - 150 dimensions, the performance of the text feature in the four machine learning algorithms overtakes the quality attributes; but when the peak is reached, quality attributes are more stable. This also provides a direction for the optimization of the quality model and the use of quality assessment in different situations

    Inter-Procedural Diagnosis Path Generation for Automatic Confirmation of Program Suspected Faults

    Get PDF
    Static analysis plays an important role in the software testing field. However, the initial results of static analysis always have a large number of false positives, which need to be confirmed by manual or automatic tools. In this paper, a novel approach is proposed, which combines the demand-driven analysis and the inter-procedural dataflow analysis, and generates the inter-procedural diagnosis paths to help the testers confirm the suspected faults automatically. In our approach, first, the influencing nodes of suspected fault are calculated. Then, the CFG of each associated procedure is simplified according to the influencing nodes. Finally, the “section-whole” strategy is employed to generate the inter-procedural diagnosis path. In order to illustrate and verify our approach, an experimental study is performed on the five open source C language projects. The results show that compared with the traditional approach, our approach requires less time and can generate more inter-procedural diagnosis paths in the given suspected faults

    Coverage Hole Recovery Algorithm Based on Molecule Model in Heterogeneous WSNs

    Get PDF
    In diverse application fields, the increasing requisitions of Wireless Sensor Networks (WSNs) have more and more research dedicated to the question of sensor nodes’ deployment in recent years. For deployment of sensor nodes, some key points that should be taken into consideration are the coverage area to be monitored, energy consumed of nodes, connectivity, amount of deployed sensors and lifetime of the WSNs. This paper analyzes the wireless sensor network nodes deployment optimization problem. Wireless sensor nodes deployment determines the nodes’ capability and lifetime. For node deployment in heterogeneous sensor networks based on different probability sensing models of heterogeneous nodes, the author refers to the organic small molecule model and proposes a molecule sensing model of heterogeneous nodes in this paper. DSmT is an extension of the classical theory of evidence, which can combine with any type of trust function of an independent source, mainly concentrating on combined uncertainty, high conflict, and inaccurate source of evidence. Referring to the data fusion model, the changes in the network coverage ratio after using the new sensing model and data fusion algorithm are studied. According to the research results, the nodes deployment scheme of heterogeneous sensor networks based on the organic small molecule model is proposed in this paper. The simulation model is established by MATLAB software. The simulation results show that the effectiveness of the algorithm, the network coverage, and detection efficiency of nodes are improved, the lifetime of the network is prolonged, energy consumption and the number of deployment nodes are reduced, and the scope of perceiving is expanded. As a result, the coverage hole recovery algorithm can improve the detection performance of the network in the initial deployment phase and coverage hole recovery phase

    Research of alarm correlations based on static defect detection

    Get PDF
    Tradicionalni alati za detekciju statičkog kvara mogu detektirati kvarove softvera i objaviti alarm, ali korelacije između alarma nisu identificirane i masivni nezavisni alarmi protivni su razumijevanju. Pomaganje korisnicima u verifikaciji alarma predstavlja veliki izazov postojećim alatima za detekciju statičke greške. U ovom radu mi formalno uvodimo korelacije alarma. Ako postojanje jednog alarma uzrokuje drugi, kažemo da su u korelaciji. Ako je jedan dominantni alarm jedinstveno povezan s drugim, znamo da će se verifikacijom jednoga također verificirati drugi. Na osnovu korelacije možemo reducirati broj alarma potrebnih za verifikaciju. Naši su algoritmi inter-proceduralni, osjetljivi na putanju i podesivi (scalable). Mi prikazujemo sumarni model postupka korelacije za računanje inter-proceduralne korelacije alarma. Osnovni algoritmi su implementirani u naše alate za detekciju kvara. Izabrali smo jednu uobičajenu semantičku pogrešku za analizu slučaja i dokazali da naša metoda rezultira smanjenjem radnog opterećenja za 34,23 %. Primjenom korelacijeske informacije možemo automatizirati verifikaciju alarma, što se ranije moralo raditi ručno.Traditional static defect detection tools can detect software defects and report alarms, but the correlations among alarms are not identified and massive independent alarms are against the understanding. Helping users in the alarm verification task is a major challenge for current static defect detection tools. In this paper, we formally introduce alarm correlations. If the occurrence of one alarm causes another alarm, we say that they are correlated. If one dominant alarm is uniquely correlated with another, we know verifying the first will also verify the others. Guided by the correlation, we can reduce the number of alarms required for verification. Our algorithms are inter-procedural, path-sensitive, and scalable. We present a correlation procedure summary model for inter-procedural alarm correlation calculation. The underlying algorithms are implemented inside our defect detection tools. We chose one common semantic fault as a case study and proved that our method has the effect of reducing 34,23 % of workload. Using correlation information, we are able to automate the alarm verification that previously had to be done manually

    AI is a viable alternative to high throughput screening: a 318-target study

    Get PDF
    : High throughput screening (HTS) is routinely used to identify bioactive small molecules. This requires physical compounds, which limits coverage of accessible chemical space. Computational approaches combined with vast on-demand chemical libraries can access far greater chemical space, provided that the predictive accuracy is sufficient to identify useful molecules. Through the largest and most diverse virtual HTS campaign reported to date, comprising 318 individual projects, we demonstrate that our AtomNet® convolutional neural network successfully finds novel hits across every major therapeutic area and protein class. We address historical limitations of computational screening by demonstrating success for target proteins without known binders, high-quality X-ray crystal structures, or manual cherry-picking of compounds. We show that the molecules selected by the AtomNet® model are novel drug-like scaffolds rather than minor modifications to known bioactive compounds. Our empirical results suggest that computational methods can substantially replace HTS as the first step of small-molecule drug discovery

    Statement-Grained Hierarchy Enhanced Code Summarization

    No full text
    Code summarization plays a vital role in aiding developers with program comprehension by generating corresponding textual descriptions for code snippets. While recent approaches have concentrated on encoding the textual and structural characteristics of source code, they often neglect the global hierarchical features, causing limited code representation. Addressing this gap, our paper introduces the statement-grained hierarchy enhanced Transformer model (SHT), a novel framework that integrates global hierarchy, syntax, and token sequences to automatically generate summaries for code snippets. SHT is distinctively designed with two encoders to learn both hierarchical and sequential features of code. One relational attention encoder processes the statement-grained hierarchical graph, producing hierarchical embeddings. Subsequently, another sequence encoder integrates these hierarchical structures with token sequences. The resulting enriched representation is then fed into a vanilla Transformer decoder, which effectively generates concise and informative summarizations. Our extensive experiments demonstrate that SHT significantly outperforms state-of-the-art approaches on two widely used Java benchmarks. This underscores the effectiveness of incorporating global hierarchical information in enhancing the quality of code summarizations

    Constructing Traceability Links between Software Requirements and Source Code Based on Neural Networks

    No full text
    Software requirement changes, code changes, software reuse, and testing are important activities in software engineering that involve the traceability links between software requirements and code. Software requirement documents, design documents, code documents, and test case documents are the intermediate products of software development. The lack of interrelationship between these documents can make it extremely difficult to change and maintain the software. Frequent requirements and code changes are inevitable in software development. Software reuse, change impact analysis, and testing also require the relationship between software requirements and code. Using these traceability links can improve the efficiency and quality of related software activities. Existing methods for constructing these links need to be better automated and accurate. To address these problems, we propose to embed software requirements and source code into feature vectors containing their semantic information based on four neural networks (NBOW, RNN, CNN, and self-attention). Accurate traceability links from requirements to code are established by comparing the similarity between these vectors. We develop a prototype tool RCT based on this method. These four networks’ performances in constructing links are explored on 18 open-source projects. The experimental results show that the self-attention network performs best, with an average Recall@50 value of 0.687 on the 18 projects, which is higher than the other three neural network models and much higher than previous approaches using information retrieval and machine learning

    Sp5 induces the expression of Nanog to maintain mouse embryonic stem cell self-renewal

    No full text
    <div><p>Activation of signal transducer and activator of transcription 3 (STAT3) by leukemia inhibitory factor (LIF) maintains mouse embryonic stem cell (mESC) self-renewal. Our previous study showed that trans-acting transcription factor 5 (Sp5), an LIF/STAT3 downstream target, supports mESC self-renewal. However, the mechanism by which Sp5 exerts these effects remains elusive. Here, we found that Nanog is a direct target of Sp5 and mediates the self-renewal-promoting effect of Sp5 in mESCs. Overexpression of <i>Sp5</i> induced <i>Nanog</i> expression, while knockdown or knockout of <i>Sp5</i> decreased the <i>Nanog</i> level. Moreover, chromatin immunoprecipitation (ChIP) assays showed that Sp5 directly bound to the Nanog promoter. Functional studies revealed that knockdown of <i>Nanog</i> eliminated the mESC self-renewal-promoting ability of Sp5. Finally, we demonstrated that the self-renewal-promoting function of Sp5 was largely dependent on its zinc finger domains. Taken together, our study provides unrecognized functions of Sp5 in mESCs and will expand our current understanding of the regulation of mESC pluripotency.</p></div

    Sp5 directly regulates the transcription of <i>Nanog</i>.

    No full text
    <p>(A) Independent validation of Nanog as an Sp5-bound target by ChIP-qPCR with fifteen primers set to scan different fragments of the Nanog promoter. Primers set at sites 1, 3, 5, 8, 9 and 15 represent significant enrichment. Data represent the mean±s.d. of three biological replicates. *p < 0.05, **p < 0.01 vs IgG. (B) Schematic illustration of luciferase reporter plasmids and <i>Sp5</i> expression activates the P<sub>Nanog</sub>-luciferase reporter. Data represent the mean±s.d. of three biological replicates. **p < 0.01 vs PB.</p
    corecore