73 research outputs found

    Cost Simulation and Performance Optimization of Web-based Applications on Mobile Channels

    Get PDF
    When considering the addition of a mobile presentation channel to an existing web-based application, a key question that has to be answered even before development begins is how the mobile channel's characteristics will impact the user experience and the cost of using the application. If either of these factors is outside acceptable limits, economical considerations may forbid adding the channels, even if it would be feasible from a purely technical perspective. Both of these factors depend considerably on two metrics: The time required to transmit data over the mobile network, and the volume transmitted. The PETTICOAT method presented in this paper uses the dialog flow model and web server log files of an existing application to identify typical interaction sequences and to compile volume statistics, which are then run through a tool that simulates the volume and time that would be incurred by executing the interaction sequences on a mobile channel. From the simulated volume and time data, we can then calculate the cost of accessing the application on a mobile channel, and derive suitable approaches for optimizing cost and response times

    An OCL-Based approach to derive constraint test cases for database applications

    Get PDF
    The development of database applications in most CASE tools has been insufficient because most of these tools do not provide the software necessary to validate these appli-cations. Validation means ensuring whether a given application fulfils the user require-ments. We suggest validation of database applications by using the functional testing technique, which is a fundamental black-box testing technique for checking the software without being concerned about its implementation and structure. Our main contribu-tion to this work is in providing a MDA approach for deriving testing software from the OCL specification of the integrity constraints. This testing software is used to validate the database applications, which are used to enforce these constraints. The generated testing software includes three components: validation queries, test cases and initial data inserted before the testing process. Our approach is implemented as an add-in tool in Rational Rose called OCL2TestSW.This work has been partially supported by the project Thuban: Natural Interaction Platform for Virtual Attending in Real Environments (TIN2008-02711), and also by the Spanish research projects: MA2VICMR: Improving the access, analysis and visibility of the multilingual and multimedia information in web for the Region of Madrid (S2009/TIC-1542).Publicad

    Guest editors’ introduction (Vol. 30, No. 7)

    Get PDF
    Editoria

    Automated analysis of feature models: Quo vadis?

    Get PDF
    Feature models have been used since the 90's to describe software product lines as a way of reusing common parts in a family of software systems. In 2010, a systematic literature review was published summarizing the advances and settling the basis of the area of Automated Analysis of Feature Models (AAFM). From then on, different studies have applied the AAFM in different domains. In this paper, we provide an overview of the evolution of this field since 2010 by performing a systematic mapping study considering 423 primary sources. We found six different variability facets where the AAFM is being applied that define the tendencies: product configuration and derivation; testing and evolution; reverse engineering; multi-model variability-analysis; variability modelling and variability-intensive systems. We also confirmed that there is a lack of industrial evidence in most of the cases. Finally, we present where and when the papers have been published and who are the authors and institutions that are contributing to the field. We observed that the maturity is proven by the increment in the number of journals published along the years as well as the diversity of conferences and workshops where papers are published. We also suggest some synergies with other areas such as cloud or mobile computing among others that can motivate further research in the future.Ministerio de EconomĂ­a y Competitividad TIN2015-70560-RJunta de AndalucĂ­a TIC-186

    An integration of train timetabling, platforming and routing-based cooperative adjustment methodology for dealing with train delay

    Get PDF
    Train delay is a serious issue that can spread rapidly in the railway network leading to further delay of other trains and detention of passengers in stations. However, the current practice in the event of the trail delay usually depends on train dispatcher's experience, which cannot manage train operation effectively and may have safety risks. The application of intelligent railway monitor and control system can improve train operation management while increasing railway safety. This paper presents a methodology in which train timetabling, platforming and routing models are combined by studying the real-time adjustment and optimization of high speed railway in the case of the train delay in order to produce a cooperative adjustment algorithm so that the train operation adjustment plan can be obtained. MATLAB computer programs have been developed based on the proposed methodology and adjustment criteria have been established from knowledge data bases in order to calculate optimized solutions. A case study is used to demonstrate the proposed methodology. The results show that the proposed method can quickly adjust the train operation plan in the case of the train delay, restore the normal train operation order, and reduce the impact of train delay on railway network effectively and efficiently

    Representation of business processes at multiple levels of abstraction (strategic, tactical and operational) during the requirements elicitation stage of a software project, and the measurement of their functional size with ISO 19761

    Get PDF
    This thesis aims at helping software engineers and business analysts to better model business processes when those models are meant to be used: for software requirements specification, and for functional size measurement purposes. The research goal of this thesis is to contribute to the representation of business processes for its use during the requirements elicitation stage of a software project. To achieve this goal, two research objectives are clearly defined: 1. To propose a novel modeling approach that generates business process models intended to be used in a software requirements elicitation activity. The modeling approach should not significantly increase the complexity of the modeling notations used to represent the business processes; and it must allow the active participation of the various stakeholders involved in a typical software project in order to represent, in a consistent and structured way, their needs and constraints. 2. To develop a procedure to measure the functional size of a software application from the business process models representing it. This measurement procedure should be compatible with the COSMIC ISO 19761 standard; and it should be able to be used independently of the modeling notation used to represent the business process. To achieve the first objective, this thesis proposes a novel modeling approach (coined BPM+) that models business processes at three levels of abstraction: strategic, tactical and operational. An a priori version of BPM+ was designed based on the findings of the literature review. This a priori version was iteratively refined through a pilot case study in industry, a series of ontological analyses, and a survey of experts. As a result, a reviewed version of BPM+ was proposed. The reviewed version was evaluated through a second case study in industry. Therefore, the design of BPM+ has been based on a triangulation of evidences obtained from various sources. To achieve the second objective, the measurement procedure was developed from an analytical comparison between the specifications of COSMIC and those of the modeling notations selected for this research (i.e. BPMN and Qualigram). This analytical comparison helped to define a set of modeling guidelines for the business application software domain. The comparison also allowed defining a set of mapping rules between the modeling notations’ constructs and the COSMIC concepts. In addition, the modeling guidelines were adapted for their application to the real-time software domain. The measurement procedure was evaluated by comparing its measurement results to those obtained in COSMIC reference case studies. The research results demonstrate that: 1. BPM+ allows generating business process models that represent in a consistent and structured way the needs of various stakeholders. 2. Qualigram notation is better suited to BPM+’s design. In addition, Qualigram notation is preferred to be used for non-IT stakeholders, while BPMN is preferred for IT stakeholders. 3. The measurement procedure was successfully applied using two different notations: Qualigram and BPMN, and in two different software domains: the business application domain and the real-time domain. 4. The accuracy of the measurement procedure is in conformity with all the rules of the ISO 19761 standard

    Evaluating and Improving Reverse Engineering Tools

    Get PDF
    Developers tend to leave some important steps and actions (e.g. properly designing the system's architecture, code review and testing) out of the software development process, and use risky practices (e.g. the copy-paste technique) so that the software can be released as fast as possible. However, these practices may turn out to be critical from the viewpoint of maintainability of the software system. In such cases, a cost-effective solution might be to re-engineer the system. Re-engineering consists of two stages, namely reverse-engineering information from the current system and, based on this information, forward-engineering the system to a new form. In this way, successful re-engineering significantly depends on the reverse engineering phase. Therefore, it is vital to guarantee correctness, and to improve the results of the reverse engineering step. Otherwise, the re-engineering of the software system could fail due to the bad results of reverse engineering. The above issues motivated us to develop a method which extends and improves one of our reverse engineering tools, and to develop benchmarks and to perform experiments on evaluating and comparing reverse engineering tools
    • 

    corecore