1,023 research outputs found
Hierarchical Control of the ATLAS Experiment
Control systems at High Energy Physics (HEP) experiments are becoming increasingly complex mainly due to the size, complexity and data volume associated to the front-end instrumentation. In particular, this becomes visible for the ATLAS experiment at the LHC accelerator at CERN. ATLAS will be the largest particle detector ever built, result of an international collaboration of more than 150 institutes. The experiment is composed of 9 different specialized sub-detectors that perform different tasks and have different requirements for operation. The system in charge of the safe and coherent operation of the whole experiment is called Detector Control System (DCS). This thesis presents the integration of the ATLAS DCS into a global control tree following the natural segmentation of the experiment into sub-detectors and smaller sub-systems. The integration of the many different systems composing the DCS includes issues such as: back-end organization, process model identification, fault detection, synchronization with external systems, automation of processes and supervisory control. Distributed control modeling is applied to the widely distributed devices that coexist in ATLAS. Thus, control is achieved by means of many distributed, autonomous and co-operative entities that are hierarchically organized and follow a finite-state machine logic. The key to integration of these systems lies in the so called Finite State Machine tool (FSM), which is based on two main enabling technologies: a SCADA product, and the State Manager Interface (SMI++) toolkit. The SMI++ toolkit has been already used with success in two previous HEP experiments providing functionality such as: an object-oriented language, a finite-state machine logic, an interface to develop expert systems, and a platform-independent communication protocol. This functionality is then used at all levels of the experiment operation process, ranging from the overall supervision down to device integration, enabling the overall sequencing and automation of the experiment. Although the experience gained in the past is an important input for the design of the detector's control hierarchy, further requirements arose due to the complexity and size of ATLAS. In total, around 200.000 channels will be supervised by the DCS and the final control tree will be hundreds of times bigger than any of the antecedents. Thus, in order to apply a hierarchical control model to the ATLAS DCS, a common approach has been proposed to ensure homogeneity between the large-scale distributed software ensembles of sub-detectors. A standard architecture and a human interface have been defined with emphasis on the early detection, monitoring and diagnosis of faults based on a dynamic fault-data mechanism. This mechanism relies on two parallel communication paths that manage the faults while providing a clear description of the detector conditions. The DCS information is split and handled by different types of SMI++ objects; whilst one path of objects manages the operational mode of the system, the other is to handle eventual faults. The proposed strategy has been validated through many different tests with positive results in both functionality and performance. This strategy has been successfully implemented and constitutes the ATLAS standard to build the global control tree. During the operation of the experiment, the DCS, responsible for the detector operation, must be synchronized with the data acquisition system which is in charge of the physics data taking process. The interaction between both systems has so far been limited, but becomes increasingly important as the detector nears completion. A prototype implementation, ready to be used during the sub-detector integration, has achieved data reconciliation by mapping the different segments of the data acquisition system into the DCS control tree. The adopted solution allows the data acquisition control applications to command different DCS sections independently and prevents incorrect physics data taking caused by a failure in a detector part. Finally, the human-machine interface presents and controls the DCS data in the ATLAS control room. The main challenges faced during the design and development phases were: how to support the operator in controlling this large system, how to maintain integration across many displays, and how to provide an effective navigation. These issues have been solved by combining the functionalities provided by both, the SCADA product and the FSM tool. The control hierarchy provides an intuitive structure for the organization of many different displays that are needed for the visualization of the experiment conditions. Each node in the tree represents a workspace that contains the functional information associated with its abstraction level within the hierarchy. By means of an effective navigation, any workspace of the control tree is accessible by the operator or detector expert within a common human interface layout. The interface is modular and flexible enough to be accommodated to new operational scenarios, fulfil the necessities of the different kind of users and facilitate the maintenance during the long lifetime of the detector of up to 20 years. The interface is in use since several months, and the sub-detector's control hierarchies, together with their associated displays, are currently being integrated into the common human-machine interface
Reverse Engineering and Testing of Rich Internet Applications
The World Wide Web experiences a continuous and constant evolution, where new initiatives, standards, approaches and technologies are continuously proposed for developing more effective and higher quality Web applications.
To satisfy the growing request of the market for Web applications, new technologies, frameworks, tools and environments that allow to develop Web and mobile applications with the least effort and in very short time have been introduced in the last years.
These new technologies have made possible the dawn of a new generation of Web applications, named Rich Internet Applications (RIAs), that offer greater usability and interactivity than traditional ones. This evolution has been accompanied by some drawbacks that are mostly due to the lack of applying well-known software engineering practices and approaches. As a consequence, new research questions and challenges have emerged in the field of web and mobile applications maintenance and testing.
The research activity described in this thesis has addressed some of these topics with the specific aim of proposing new and effective solutions to the problems of modelling, reverse engineering, comprehending, re-documenting and testing existing RIAs.
Due to the growing relevance of mobile applications in the renewed Web scenarios, the problem of testing mobile applications developed for the Android operating system has been addressed too, in an attempt of exploring and proposing new techniques of testing automation for these type of applications
A Brief History of Web Crawlers
Web crawlers visit internet applications, collect data, and learn about new
web pages from visited pages. Web crawlers have a long and interesting history.
Early web crawlers collected statistics about the web. In addition to
collecting statistics about the web and indexing the applications for search
engines, modern crawlers can be used to perform accessibility and vulnerability
checks on the application. Quick expansion of the web, and the complexity added
to web applications have made the process of crawling a very challenging one.
Throughout the history of web crawling many researchers and industrial groups
addressed different issues and challenges that web crawlers face. Different
solutions have been proposed to reduce the time and cost of crawling.
Performing an exhaustive crawl is a challenging question. Additionally
capturing the model of a modern web application and extracting data from it
automatically is another open question. What follows is a brief history of
different technique and algorithms used from the early days of crawling up to
the recent days. We introduce criteria to evaluate the relative performance of
web crawlers. Based on these criteria we plot the evolution of web crawlers and
compare their performanc
User Feedback-Informed Interface Design for Flow Management Data and Services (FMDS)
The transition to a microservices-based Flow Management Data and Services
(FMDS) architecture from the existing Traffic Flow Management System (TFMS) is
a critical enabler of the vision for an Information-Centric National Airspace
System (NAS). The need to design a user-centric interface for FMDS is a key
technical gap, as this interface connects NAS data and services to the traffic
management specialists within all stakeholder groups (e.g., FAA, airlines). We
provide a research-driven approach towards designing such a graphical user
interface (GUI) for FMDS. Major goals include unifying the more than 50
disparate traffic management services currently hosted on TFMS, as well as
streamlining the process of evaluating, modeling, and monitoring Traffic
Management Initiatives (TMIs). Motivated by this, we iteratively designed a GUI
leveraging human factors engineering and user experience design principles, as
well as user interviews. Through user testing and interviews, we identify
workflow benefits of our GUI (e.g., reduction in task completion time), along
with next steps for developing a live prototype.Comment: 8 pages, 8 figure
Dynamic Context Modeling for Agile Case Management
International audienceCase Management processes are characterized by their high unpredictability and, thus, cannot be handled following traditional process- or activity-centered approaches. Adaptive Case Management paradigm proposes an alternative data-centered approach for management such processes. In this paper, we elaborate on this approach and explore the role of context data in Case Management. We use the state-oriented representation of the process that allows us to incorporate the contextual information in a systematic and transparent way, leading towards agile case management
Functional Size Measurement and Model Verification for Software Model-Driven Developments: A COSMIC-based Approach
Historically, software production methods and tools have a unique goal: to produce high quality
software. Since the goal of Model-Driven Development (MDD) methods is no different, MDD
methods have emerged to take advantage of the benefits of using conceptual models to produce
high quality software.
In such MDD contexts, conceptual models are used as input to automatically generate final
applications. Thus, we advocate that there is a relation between the quality of the final software
product and the quality of the models used to generate it. The quality of conceptual models can
be influenced by many factors. In this thesis, we focus on the accuracy of the techniques used to
predict the characteristics of the development process and the generated products.
In terms of the prediction techniques for software development processes, it is widely
accepted that knowing the functional size of applications in order to successfully apply effort
models and budget models is essential. In order to evaluate the quality of generated
applications, defect detection is considered to be the most suitable technique.
The research goal of this thesis is to provide an accurate measurement procedure based on
COSMIC for the automatic sizing of object-oriented OO-Method MDD applications. To
achieve this research goal, it is necessary to accurately measure the conceptual models used in
the generation of object-oriented applications. It is also very important for these models not to
have defects so that the applications to be measured are correctly represented.
In this thesis, we present the OOmCFP (OO-Method COSMIC Function Points) measurement
procedure. This procedure makes a twofold contribution: the accurate measurement of objectoriented
applications generated in MDD environments from the conceptual models involved, and
the verification of conceptual models to allow the complete generation of correct final applications
from the conceptual models involved.
The OOmCFP procedure has been systematically designed, applied, and
automated. This measurement procedure has been validated to conform to the
ISO 14143 standard, the metrology concepts defined in the ISO VIM, and the
accuracy of the measurements obtained according to ISO 5725. This
procedure has also been validated by performing empirical studies.
The results of the empirical studies demonstrate that OOmCFP can obtain
accurate measures of the functional size of applications generated in MDD
environments from the corresponding conceptual models.MarÃn Campusano, BM. (2011). Functional Size Measurement and Model Verification for Software Model-Driven Developments: A COSMIC-based Approach [Tesis doctoral no publicada]. Universitat Politècnica de València. https://doi.org/10.4995/Thesis/10251/11237Palanci
Automated specification-based testing of graphical user interfaces
Tese de doutoramento. Engenharia Electrónica e de Computadores. 2006. Faculdade de Engenharia. Universidade do Porto, Departamento de Informática, Escola de Engenharia. Universidade do Minh
- …