8 research outputs found

    Pragmatic cost estimation for web applications

    Get PDF
    Cost estimation for web applications is an interesting and difficult challenge for researchers and industrial practitioners. It is a particularly valuable area of ongoing commercial research. Attaining on accurate cost estimation for web applications is an essential element in being able to provide competitive bids and remaining successful in the market. The development of prediction techniques over thirty years ago has contributed to several different strategies. Unfortunately there is no collective evidence to give substantial advice or guidance for industrial practitioners. Therefore to address this problem, this thesis shows the way by investigating the characteristics of the dataset by combining the literature review and industrial survey findings. The results of the systematic literature review, industrial survey and an initial investigation, have led to an understanding that dataset characteristics may influence the cost estimation prediction techniques. From this, an investigation was carried out on dataset characteristics. However, in the attempt to structure the characteristics of dataset it was found not to be practical or easy to get a defined structure of dataset characteristics to use as a basis for prediction model selection. Therefore the thesis develops a pragmatic cost estimation strategy based on collected advice and general sound practice in cost estimation. The strategy is composed of the following five steps: test whether the predictions are better than the means of the dataset; test the predictions using accuracy measures such as MMRE, Pred and MAE knowing their strengths and weaknesses; investigate the prediction models formed to see if they are sensible and reasonable model; perform significance testing on the predictions; and get the effect size to establish preference relations of prediction models. The results from this pragmatic cost estimation strategy give not only advice on several techniques to choose from, but also give reliable results. Practitioners can be more confident about the estimation that is given by following this pragmatic cost estimation strategy. It can be concluded that the practitioners should focus on the best strategy to apply in cost estimation rather than focusing on the best techniques. Therefore, this pragmatic cost estimation strategy could help researchers and practitioners to get reliable results. The improvement and replication of this strategy over time will produce much more useful and trusted results.Cost estimation for web applications is an interesting and difficult challenge for researchers and industrial practitioners. It is a particularly valuable area of ongoing commercial research. Attaining on accurate cost estimation for web applications is an essential element in being able to provide competitive bids and remaining successful in the market. The development of prediction techniques over thirty years ago has contributed to several different strategies. Unfortunately there is no collective evidence to give substantial advice or guidance for industrial practitioners. Therefore to address this problem, this thesis shows the way by investigating the characteristics of the dataset by combining the literature review and industrial survey findings. The results of the systematic literature review, industrial survey and an initial investigation, have led to an understanding that dataset characteristics may influence the cost estimation prediction techniques. From this, an investigation was carried out on dataset characteristics. However, in the attempt to structure the characteristics of dataset it was found not to be practical or easy to get a defined structure of dataset characteristics to use as a basis for prediction model selection. Therefore the thesis develops a pragmatic cost estimation strategy based on collected advice and general sound practice in cost estimation. The strategy is composed of the following five steps: test whether the predictions are better than the means of the dataset; test the predictions using accuracy measures such as MMRE, Pred and MAE knowing their strengths and weaknesses; investigate the prediction models formed to see if they are sensible and reasonable model; perform significance testing on the predictions; and get the effect size to establish preference relations of prediction models. The results from this pragmatic cost estimation strategy give not only advice on several techniques to choose from, but also give reliable results. Practitioners can be more confident about the estimation that is given by following this pragmatic cost estimation strategy. It can be concluded that the practitioners should focus on the best strategy to apply in cost estimation rather than focusing on the best techniques. Therefore, this pragmatic cost estimation strategy could help researchers and practitioners to get reliable results. The improvement and replication of this strategy over time will produce much more useful and trusted results

    Topic Modeling in Sentiment Analysis: A Systematic Review

    Get PDF
    With the expansion and acceptance of Word Wide Web, sentiment analysis has become progressively popular research area in information retrieval and web data analysis. Due to the huge amount of user-generated contents over blogs, forums, social media, etc., sentiment analysis has attracted researchers both in academia and industry, since it deals with the extraction of opinions and sentiments. In this paper, we have presented a review of topic modeling, especially LDA-based techniques, in sentiment analysis. We have presented a detailed analysis of diverse approaches and techniques, and compared the accuracy of different systems among them. The results of different approaches have been summarized, analyzed and presented in a sophisticated fashion. This is the really effort to explore different topic modeling techniques in the capacity of sentiment analysis and imparting a comprehensive comparison among them

    Software quality: predicting reliability of a software using decision tree

    Get PDF
    System availability can be expressed as an attribute of reliability that determines the total time a system or component is functioning. Most available models try to predict availability of a software during its life cycle but there are very few or no models that predict a software going days without a failure. Over the years, decision tree model have been used as a reliable technique for prediction. In this study, based on the sample data collected by John Musa of Bell Telephone Laboratories, a decision tree model has been used to predict the availability of a system going days without a failure. This study concluded that a decision tree model is able to decide availability of a software in terms of going days without a failure

    Predictive analytic in health care using Case-based Reasoning (CBR)

    Get PDF
    Big data analytics enables useful information to be extracted in order to predict trends and behavior patterns.Predictive analytics can be applied in health care industry by using the information gained from big data analytics.There are several methods to make predictive analytics. Casebased Reasoning (CBR) is one of the methods to make prediction on patients’ sickness based on previous experiences.There are several challenges when applying CBR to predictive analytics.This paper focuses on solving the number of analogies used when applying CBR.Experiments and calculations are done to compare the accuracy of the number of analogies used.The results shows one analogy has the highest accuracy as compared to two and three analogies

    Spatio-temporal crime HotSpot detection and prediction: a systematic literature review

    Get PDF
    The primary objective of this study is to accumulate, summarize, and evaluate the state-of-the-art for spatio-temporal crime hotspot detection and prediction techniques by conducting a systematic literature review (SLR). The authors were unable to find a comprehensive study on crime hotspot detection and prediction while conducting this SLR. Therefore, to the best of author's knowledge, this study is the premier attempt to critically analyze the existing literature along with presenting potential challenges faced by current crime hotspot detection and prediction systems. The SLR is conducted by thoroughly consulting top five scientific databases (such as IEEE, Science Direct, Springer, Scopus, and ACM), and synthesized 49 different studies on crime hotspot detection and prediction after critical review. This study unfolds the following major aspects: 1) the impact of data mining and machine learning approaches, especially clustering techniques in crime hotspot detection; 2) the utility of time series analysis techniques and deep learning techniques in crime trend prediction; 3) the inclusion of spatial and temporal information in crime datasets making the crime prediction systems more accurate and reliable; 4) the potential challenges faced by the state-of-the-art techniques and the future research directions. Moreover, the SLR aims to provide a core foundation for the research on spatio-temporal crime prediction applications while highlighting several challenges related to the accuracy of crime hotspot detection and prediction applications

    Spatio-temporal crime predictions by leveraging artificial intelligence for citizens security in smart cities

    Get PDF
    Smart city infrastructure has a significant impact on improving the quality of humans life. However, a substantial increase in the urban population from the last few years poses challenges related to resource management, safety, and security. To ensure the safety and security in the smart city environment, this paper presents a novel approach by empowering the authorities to better visualize the threats, by identifying and predicting the highly-reported crime zones in the smart city. To this end, it first investigates the Hierarchical Density-Based Spatial Clustering of Applications with Noise (HDBSCAN) to detect the hot-spots that have a higher risk of crime occurrence. Second, for crime prediction, Seasonal Auto-Regressive Integrated Moving Average (SARIMA) is exploited in each dense crime region to predict the number of crime incidents in the future with spatial and temporal information. The proposed HDBSCAN and SARIMA based crime prediction model is evaluated on ten years of crime data (2008-2017) for New York City (NYC) . The accuracy of the model is measured by considering different time scenarios such as the year-wise, (i.e., for each year), and for the total considered duration of ten years using an 80:20 ratio. The 80% of data was used for training and 20% for testing. The proposed approach outperforms with an average Mean Absolute Error (MAE) of 11.47 as compared to the highest scoring DBSCAN based method with MAE 27.03

    Usability measurement of Malaysian online tourism websites

    No full text
    Tourism in Malaysia has huge potential as the country is rich with different kinds of attractions such as historical places, different kinds of festival to enjoy and beautiful scenery to embrace. With the advances of the Internet,many investors have begun to produce different websites to promote their product; however, the quality is still valued expeditiously. The purpose of this paper is to analyze and evaluate the quality of Malaysian tourism websites in terms of usability. With this aim, five different Malaysian tourism websites were selected and evaluated based on their effectiveness, efficiency and user satisfaction. This was done using a questionnaire based evaluation

    Deep Learning-Based Growth Prediction System: A Use Case of China Agriculture

    No full text
    Agricultural advancements have significantly impacted people’s lives and their surroundings in recent years. The insufficient knowledge of the whole agricultural production system and conventional ways of irrigation have limited agricultural yields in the past. The remote sensing innovations recently implemented in agriculture have dramatically revolutionized production efficiency by offering unparalleled opportunities for convenient, versatile, and quick collection of land images to collect critical details on the crop’s conditions. These innovations have enabled automated data collection, simulation, and interpretation based on crop analytics facilitated by deep learning techniques. This paper aims to reveal the transformative patterns of old Chinese agrarian development and fruit production by focusing on the major crop production (from 1980 to 2050) taking into account various forms of data from fruit production (e.g., apples, bananas, citrus fruits, pears, and grapes). In this study, we used production data for different fruits grown in China to predict the future production of these fruits. The study employs deep neural networks to project future fruit production based on the statistics issued by China’s National Bureau of Statistics on the total fruit growth output for this period. The proposed method exhibits encouraging results with an accuracy of 95.56% calculating by accuracy formula based on fruit production variation. Authors further provide recommendations on the AGR-DL (agricultural deep learning) method being helpful for developing countries. The results suggest that the agricultural development in China is acceptable but demands more improvement and government needs to prioritize expanding the fruit production by establishing new strategies for cultivators to boost their performance
    corecore