1,701 research outputs found

    Calibrating Function Point Backfiring Conversion Ratios Using Neuro-Fuzzy Technique

    Get PDF
    Software estimation is an important aspect in software development projects because poor estimations can lead to late delivery, cost overruns, and possibly project failure. Backfiring is a popular technique for sizing and predicting the volume of source code by converting the function point metric into source lines of code mathematically using conversion ratios. While this technique is popular and useful, there is a high margin of error in backfiring. This research introduces a new method to reduce that margin of error. Neural networks and fuzzy logic in software prediction models have been demonstrated in the past to have improved performance over traditional techniques. For this reason, a neuro-fuzzy approach is introduced to the backfiring technique to calibrate the conversion ratios. This paper presents the neuro-fuzzy calibration solution and compares the calibrated model against the default conversion ratios currently used by software practitioners

    A particle swarm optimisation-based Grey prediction model for thermal error compensation on CNC machine tools

    Get PDF
    Thermal errors can have a significant effect on CNC machine tool accuracy. The thermal error compensation system has become a cost-effective method of improving machine tool accuracy in recent years. In the presented paper, the Grey relational analysis (GRA) was employed to obtain the similarity degrees between fixed temperature sensors and the thermal response of the CNC machine tool structure. Subsequently, a new Grey model with convolution integral GMC(1, N) is used to design a thermal prediction model. To improve the accuracy of the proposed model, the generation coefficients of GMC(1, N) are calibrated using an adaptive Particle Swarm Optimisation (PSO) algorithm. The results demonstrate good agreement between the experimental and predicted thermal error. Finally, the capabilities and the limitations of the model for thermal error compensation have been discussed. Keywords: CNC machine tool, Thermal error modelling, ANFIS, Fuzzy logic, Grey system theory

    A New Calibration for Function Point Complexity Weights

    Get PDF
    Function Point (FP) is a useful software metric that was first proposed twenty-five years ago, since then, it has steadily evolved into a functional size metric consolidated in the well-accepted Standardized International Function Point Users Group (IFPUG) Counting Practices Manual - version 4.2. While software development industry has grown rapidly, the weight values assigned to count standard FP still remain same, which raise critical questions about the validity of the weight values. In this paper, we discuss the concepts of calibrating Function Point, whose aims are to estimate a more accurate software size that fits for specific software application, to reflect software industry trend, and to improve the cost estimation of software projects. A FP calibration model called Neuro-Fuzzy Function Point Calibration Model (NFFPCM) that integrates the learning ability from neural network and the ability to capture human knowledge from fuzzy logic is proposed. The empirical validation using International Software Benchmarking Standards Group (ISBSG) data repository release 8 shows a 22% accuracy improvement of mean MRE in software effort estimation after calibration

    A Soft Computing Framework for Software Effort Estimation

    Get PDF
    Accurate software estimation such as cost estimation, quality estimation and risk analysis is a major issue in software project management. In this paper, we present a soft computing framework to tackle this challenging problem. We first use a preprocessing neuro-fuzzy inference system to handle the dependencies among contributing factors and decouple the effects of the contributing factors into individuals. Then we use a neuro-fuzzy bank to calibrate the parameters of contributing factors. In order to extend our framework into fields that lack of an appropriate algorithmic model of their own, we propose a default algorithmic model that can be replaced when a better model is available. One feature of this framework is that the architecture is inherently independent of the choice of algorithmic models or the nature of the estimation problems. By integrating neural networks, fuzzy logic and algorithmic models into one scheme, this framework has learning ability, integration capability of both expert knowledge and project data, good interpretability, and robustness to imprecise and uncertain inputs. Validation using industry project data shows that the framework produces good results when used to predict software cost

    AI and OR in management of operations: history and trends

    Get PDF
    The last decade has seen a considerable growth in the use of Artificial Intelligence (AI) for operations management with the aim of finding solutions to problems that are increasing in complexity and scale. This paper begins by setting the context for the survey through a historical perspective of OR and AI. An extensive survey of applications of AI techniques for operations management, covering a total of over 1200 papers published from 1995 to 2004 is then presented. The survey utilizes Elsevier's ScienceDirect database as a source. Hence, the survey may not cover all the relevant journals but includes a sufficiently wide range of publications to make it representative of the research in the field. The papers are categorized into four areas of operations management: (a) design, (b) scheduling, (c) process planning and control and (d) quality, maintenance and fault diagnosis. Each of the four areas is categorized in terms of the AI techniques used: genetic algorithms, case-based reasoning, knowledge-based systems, fuzzy logic and hybrid techniques. The trends over the last decade are identified, discussed with respect to expected trends and directions for future work suggested
    corecore