268,245 research outputs found
Virtual learning process environment (VLPE): a BPM-based learning process management architecture
E-learning systems have signiïŹcantly impacted the way that learning takes place within universities, particularly in providing self-learning support and ïŹexibility of course delivery. Virtual Learning Environments help facilitate the management of educational courses for students, in particular by assisting course designers and thriving in the management of the learning itself. Current literature has shown that pedagogical modelling and learning process management facilitation are inadequate. In particular, quantitative information on the process of learning that is needed to perform real time or reïŹective monitoring and statistical analysis of studentsâ learning processes performance is deïŹcient. Therefore, for a course designer, pedagogical evaluation and reform decisions can be diïŹcult. This thesis presents an alternative e-learning systems architecture - Virtual Learning Process Environment (VLPE) - that uses the Business Process Management (BPM) conceptual framework to design an architecture that addresses the critical quantitative learning process information gaps associated with the conventional VLE frameworks. Within VLPE, course designers can model desired education pedagogies in the form of learning process workïŹows using an intuitive graphical ïŹow diagram user-interface. Automated agents associated with BPM frameworks are employed to capture quantitative learning information from the learning process workïŹow. Consequently, course designers are able to monitor, analyse and re-evaluate in real time the eïŹectiveness of their chosen pedagogy using live interactive learning process dashboards. Once a course delivery is complete the collated quantitative information can also be used to make major revisions to pedagogy design for the next iteration of the course. An additional contribution of this work is that this new architecture facilitates individual students in monitoring and analysing their own learning performances in comparison to their peers in a real time anonymous manner through a personal analytics learning process dashboard. A case scenario of the quantitative statistical analysis of a cohort of learners (10 participants in size) is presented. The analytical results of their learning processes, performances and progressions on a short Mathematics course over a ïŹve-week period are also presented in order to demonstrate that the proposed framework can signiïŹcantly help to advance learning analytics and the visualisation of real time learning data
Monitoring of Spatial Data Infraestructures
SDI monitoring and evaluation is increasingly attracting the attention of both public sector bureaucrats seeking justification for providing public sources to SDI and SDI practitioners requiring a measure of success of their SDI strategy. In recent years, a shift from an intuitive to more rational SDI assessments can be observed. SDI monitoring and evaluation is becoming operational and is already part of some SDI implementations and practices. Based on an analysis of the operational monitoring systems of the Dutch national SDI (GIDEON), the European SDI (INSPIRE) and the Catalan SDI (IDEC). We describe, analyze and compare comprehensively the design and application of operational SDI monitoring systems and identify common issues to be taken into account for monitoring of SDIs. This can support further improvement of evaluation practices and operational setups of SDI monitoring systems
Report from GI-Dagstuhl Seminar 16394: Software Performance Engineering in the DevOps World
This report documents the program and the outcomes of GI-Dagstuhl Seminar
16394 "Software Performance Engineering in the DevOps World".
The seminar addressed the problem of performance-aware DevOps. Both, DevOps
and performance engineering have been growing trends over the past one to two
years, in no small part due to the rise in importance of identifying
performance anomalies in the operations (Ops) of cloud and big data systems and
feeding these back to the development (Dev). However, so far, the research
community has treated software engineering, performance engineering, and cloud
computing mostly as individual research areas. We aimed to identify
cross-community collaboration, and to set the path for long-lasting
collaborations towards performance-aware DevOps.
The main goal of the seminar was to bring together young researchers (PhD
students in a later stage of their PhD, as well as PostDocs or Junior
Professors) in the areas of (i) software engineering, (ii) performance
engineering, and (iii) cloud computing and big data to present their current
research projects, to exchange experience and expertise, to discuss research
challenges, and to develop ideas for future collaborations
Adaptive planning for distributed systems using goal accomplishment tracking
Goal accomplishment tracking is the process of monitoring the progress of a task or series of tasks towards completing a goal. Goal accomplishment tracking is used to monitor goal progress in a variety of domains, including workflow processing, teleoperation and industrial manufacturing. Practically, it involves the constant monitoring of task execution, analysis of this data to determine the task progress and notification of interested parties. This information is usually used in a passive way to observe goal progress. However, responding to this information may prevent goal failures. In addition, responding proactively in an opportunistic way can also lead to goals being completed faster. This paper proposes an architecture to support the adaptive planning of tasks for fault tolerance or opportunistic task execution based on goal accomplishment tracking. It argues that dramatically increased performance can be gained by monitoring task execution and altering plans dynamically
Service Knowledge Capture and Reuse
The keynote will start with the need for service knowledge capture and reuse for industrial product-service systems. A novel approach to capture the service damage knowledge about individual component will be presented with experimental results. The technique uses active thermography and image processing approaches for the assessment. The paper will also give an overview of other non-destructive inspection techniques for service damage assessment. A robotic system will be described to automate the damage image capture. The keynote will then propose ways to reuse the knowledge to predict remaining life of the component and feedback to design and manufacturing
Synergistic combination of systems for structural health monitoring and earthquake early warning for structural health prognosis and diagnosis
Earthquake early warning (EEW) systems are currently operating nationwide in Japan and are in beta-testing in California. Such a system detects an earthquake initiation using online signals from a seismic sensor network and broadcasts a warning of the predicted location and magnitude a few seconds to a minute or so before an earthquake hits a site. Such a system can be used synergistically with installed structural health monitoring (SHM) systems to enhance pre-event prognosis and post-event diagnosis of structural health. For pre-event prognosis, the EEW system information can be used to make probabilistic predictions of the anticipated damage to a structure using seismic loss estimation methodologies from performance-based earthquake engineering. These predictions can support decision-making regarding the activation of appropriate mitigation systems, such as stopping traffic from entering a bridge that has a predicted high probability of damage. Since the time between warning and arrival of the strong shaking is very short, probabilistic predictions must be rapidly calculated and the decision making automated for the mitigation actions. For post-event diagnosis, the SHM sensor data can be used in Bayesian updating of the probabilistic damage predictions with the EEW predictions as a prior. Appropriate Bayesian methods for SHM have been published. In this paper, we use pre-trained surrogate models (or emulators) based on machine learning methods to make fast damage and loss predictions that are then used in a cost-benefit decision framework for activation of a mitigation measure. A simple illustrative example of an infrastructure application is presented
- âŠ