704,116 research outputs found
Recommended from our members
Alignment of IT projects with business strategy: An analysis of the interrelationships between the factors affecting IS alignment at strategic, tactical and operational levels
This thesis was submitted for the degree of Doctor of Philosophy and awarded by Brunel University.Despite numerous efforts to integrate business and IS strategic plans, organisations are
not delivering the expected benefits from IS investment. To address this issue, IS
alignment research has discussed extensively the idea of establishing two-way commitment between business and IS managers. This commitment, however, has proved to be difficult to achieve at strategic level and consequently difficult to transmit to lower levels within organisations. Given that current literature has identified the main factors affecting IS alignment, this research extends the analysis of those factors to tactical and operational levels to develop a model that depicts the dynamic interrelationships between the factors affecting IS alignment. Through an interpretative approach that combines quantitative and qualitative methods, the model was developed, tested and evaluated in three phases.
During the exploratory phase the factors IT governance, communication, partnership, IT
value, scope & architecture and human resources skills were scrutinised using a pilot case
study and a survey. The results aided the selection of relevant variables that could be used
in the model to assess alignment across different levels, and therefore, to develop a preliminary model that included the initial relationships between the factors. For the testing phase, a case study approach was selected. An IS alignment assessment process was designed and applied in one SME and one large organisation. Although the
assessment process did not prove appropriate in an SME context, the application of the assessment process in the large organisation allowed the identification of the root causes of high or low levels of IS alignment of five strategic IT projects. For the evaluation phase further analysis was conducted to modify the preliminary model in the light of the outcomes from the large organisation.
The findings from the evaluation phase helped in the identification of two categories of factors (structural and dynamic) and how they interrelate, and these are incorporated into the final model. Structural factors refer to those cultural and structural forces that determine whether the information systems function is valued or not as a partner in delivering business value from IT investments. On the other hand, the dynamic factors refer to those aspects that impact on IS alignment as a result of the dynamic interaction between the people involved in the strategy formulation and implementation. The model
and the assessment process represent a contribution towards a better understanding of the
nature of IS alignment
Living Innovation Laboratory Model Design and Implementation
Living Innovation Laboratory (LIL) is an open and recyclable way for
multidisciplinary researchers to remote control resources and co-develop user
centered projects. In the past few years, there were several papers about LIL
published and trying to discuss and define the model and architecture of LIL.
People all acknowledge about the three characteristics of LIL: user centered,
co-creation, and context aware, which make it distinguished from test platform
and other innovation approaches. Its existing model consists of five phases:
initialization, preparation, formation, development, and evaluation.
Goal Net is a goal-oriented methodology to formularize a progress. In this
thesis, Goal Net is adopted to subtract a detailed and systemic methodology for
LIL. LIL Goal Net Model breaks the five phases of LIL into more detailed steps.
Big data, crowd sourcing, crowd funding and crowd testing take place in
suitable steps to realize UUI, MCC and PCA throughout the innovation process in
LIL 2.0. It would become a guideline for any company or organization to develop
a project in the form of an LIL 2.0 project.
To prove the feasibility of LIL Goal Net Model, it was applied to two real
cases. One project is a Kinect game and the other one is an Internet product.
They were both transformed to LIL 2.0 successfully, based on LIL goal net based
methodology. The two projects were evaluated by phenomenography, which was a
qualitative research method to study human experiences and their relations in
hope of finding the better way to improve human experiences. Through
phenomenographic study, the positive evaluation results showed that the new
generation of LIL had more advantages in terms of effectiveness and efficiency.Comment: This is a book draf
Maintenance of Automated Test Suites in Industry: An Empirical study on Visual GUI Testing
Context: Verification and validation (V&V) activities make up 20 to 50
percent of the total development costs of a software system in practice. Test
automation is proposed to lower these V&V costs but available research only
provides limited empirical data from industrial practice about the maintenance
costs of automated tests and what factors affect these costs. In particular,
these costs and factors are unknown for automated GUI-based testing.
Objective: This paper addresses this lack of knowledge through analysis of
the costs and factors associated with the maintenance of automated GUI-based
tests in industrial practice.
Method: An empirical study at two companies, Siemens and Saab, is reported
where interviews about, and empirical work with, Visual GUI Testing is
performed to acquire data about the technique's maintenance costs and
feasibility.
Results: 13 factors are observed that affect maintenance, e.g. tester
knowledge/experience and test case complexity. Further, statistical analysis
shows that developing new test scripts is costlier than maintenance but also
that frequent maintenance is less costly than infrequent, big bang maintenance.
In addition a cost model, based on previous work, is presented that estimates
the time to positive return on investment (ROI) of test automation compared to
manual testing.
Conclusions: It is concluded that test automation can lower overall software
development costs of a project whilst also having positive effects on software
quality. However, maintenance costs can still be considerable and the less time
a company currently spends on manual testing, the more time is required before
positive, economic, ROI is reached after automation
Incorporating Agile with MDA Case Study: Online Polling System
Nowadays agile software development is used in greater extend but for small
organizations only, whereas MDA is suitable for large organizations but yet not
standardized. In this paper the pros and cons of Model Driven Architecture
(MDA) and Extreme programming have been discussed. As both of them have some
limitations and cannot be used in both large scale and small scale
organizations a new architecture has been proposed. In this model it is tried
to opt the advantages and important values to overcome the limitations of both
the software development procedures. In support to the proposed architecture
the implementation of it on Online Polling System has been discussed and all
the phases of software development have been explained.Comment: 14 pages,1 Figure,1 Tabl
A Model-Driven Approach for Business Process Management
The Business Process Management is a common mechanism recommended by a high number of standards for the management of companies and organizations. In software companies this practice is every day more accepted and companies have to assume it, if they want to be competitive. However, the effective definition of these processes and mainly their maintenance and execution are not always easy tasks. This paper presents an approach based on the Model-Driven paradigm for Business Process Management in software companies. This solution offers a suitable mechanism that was implemented successfully in different companies with a tool case named NDTQ-Framework.Ministerio de Educación y Ciencia TIN2010-20057-C03-02Junta de Andalucía TIC-578
Data Minimisation in Communication Protocols: A Formal Analysis Framework and Application to Identity Management
With the growing amount of personal information exchanged over the Internet,
privacy is becoming more and more a concern for users. One of the key
principles in protecting privacy is data minimisation. This principle requires
that only the minimum amount of information necessary to accomplish a certain
goal is collected and processed. "Privacy-enhancing" communication protocols
have been proposed to guarantee data minimisation in a wide range of
applications. However, currently there is no satisfactory way to assess and
compare the privacy they offer in a precise way: existing analyses are either
too informal and high-level, or specific for one particular system. In this
work, we propose a general formal framework to analyse and compare
communication protocols with respect to privacy by data minimisation. Privacy
requirements are formalised independent of a particular protocol in terms of
the knowledge of (coalitions of) actors in a three-layer model of personal
information. These requirements are then verified automatically for particular
protocols by computing this knowledge from a description of their
communication. We validate our framework in an identity management (IdM) case
study. As IdM systems are used more and more to satisfy the increasing need for
reliable on-line identification and authentication, privacy is becoming an
increasingly critical issue. We use our framework to analyse and compare four
identity management systems. Finally, we discuss the completeness and
(re)usability of the proposed framework
An integrated search-based approach for automatic testing from extended finite state machine (EFSM) models
This is the post-print version of the Article - Copyright @ 2011 ElsevierThe extended finite state machine (EFSM) is a modelling approach that has been used to represent a wide range of systems. When testing from an EFSM, it is normal to use a test criterion such as transition coverage. Such test criteria are often expressed in terms of transition paths (TPs) through an EFSM. Despite the popularity of EFSMs, testing from an EFSM is difficult for two main reasons: path feasibility and path input sequence generation. The path feasibility problem concerns generating paths that are feasible whereas the path input sequence generation problem is to find an input sequence that can traverse a feasible path. While search-based approaches have been used in test automation, there has been relatively little work that uses them when testing from an EFSM. In this paper, we propose an integrated search-based approach to automate testing from an EFSM. The approach has two phases, the aim of the first phase being to produce a feasible TP (FTP) while the second phase searches for an input sequence to trigger this TP. The first phase uses a Genetic Algorithm whose fitness function is a TP feasibility metric based on dataflow dependence. The second phase uses a Genetic Algorithm whose fitness function is based on a combination of a branch distance function and approach level. Experimental results using five EFSMs found the first phase to be effective in generating FTPs with a success rate of approximately 96.6%. Furthermore, the proposed input sequence generator could trigger all the generated feasible TPs (success rate = 100%). The results derived from the experiment demonstrate that the proposed approach is effective in automating testing from an EFSM
A Model-Driven approach for functional test case generation
Test phase is one of the most critical phases in software engineering life cycle to assure the final system quality. In this context, functional system test cases verify that the system under test fulfills its functional specification. Thus, these test cases are frequently designed from the different scenarios and alternatives depicted in functional requirements. The objective of this paper is to introduce a systematic process based on the Model-Driven paradigm to automate the generation of functional test cases from functional requirements. For this aim, a set of metamodels and transformations and also a specific language domain to use them is presented. The paper finishes stating learned lessons from the trenches as well as relevant future work and conclusions that draw new research lines in the test cases generation context.Ministerio de Economía y Competitividad TIN2013-46928-C3-3-
Robustness-Driven Resilience Evaluation of Self-Adaptive Software Systems
An increasingly important requirement for certain classes of software-intensive systems is the ability to self-adapt their structure and behavior at run-time when reacting to changes that may occur to the system, its environment, or its goals. A major challenge related to self-adaptive software systems is the ability to provide assurances of their resilience when facing changes. Since in these systems, the components that act as controllers of a target system incorporate highly complex software, there is the need to analyze the impact that controller failures might have on the services delivered by the system. In this paper, we present a novel approach for evaluating the resilience of self-adaptive software systems by applying robustness testing techniques to the controller to uncover failures that can affect system resilience. The approach for evaluating resilience, which is based on probabilistic model checking, quantifies the probability of satisfaction of system properties when the target system is subject to controller failures. The feasibility of the proposed approach is evaluated in the context of an industrial middleware system used to monitor and manage highly populated networks of devices, which was implemented using the Rainbow framework for architecture-based self-adaptation
- …