136 research outputs found
COLEG: Collaborative Learning Environment within Grid
The principal function of the CSCL environments is to provide to the various users (students, teachers, tutors…), the best activities with the best tools at the best time according to their needs. If a CSCL system is a collection of activities or learning process, we can cut out its functionalities in a certain number of autonomous functions which can then be carried out separately in the form of autonomous applications by using the technology of the Web/Grid services. The emerging technologies based on the Grid are increasingly being adopted to improve education and provide better services for learning. These services are offered to students who, regardless of their computer systems, can collaborate to improve their cognitive and social skills. This article presents COLEG (COllaborative Learning Environment within Grid), which aims to employ the capacities offered by the Grid to give the various actors, all the power of learning, collaboration and communication in an adaptable, heterogeneous and dynamic sight
Interim research assessment 2003-2005 - Computer Science
This report primarily serves as a source of information for the 2007 Interim Research Assessment Committee for Computer Science at the three technical universities in the Netherlands. The report also provides information for others interested in our research activities
A reactive architecture for cloud-based system engineering
PhD ThesisSoftware system engineering is increasingly practised over globally distributed locations. Such a practise is termed as Global Software Development (GSD). GSD has become a business necessity mainly because of the
scarcity of resources, cost, and the need to locate development closer to
the customers. GSD is highly dependent on requirements management,
but system requirements continuously change. Poorly managed change in
requirements affects the overall cost, schedule and quality of GSD projects.
It is particularly challenging to manage and trace such changes, and hence
we require a rigorous requirement change management (RCM) process.
RCM is not trivial in collocated software development; and with the presence of geographical, cultural, social and temporal factors, it makes RCM
profoundly difficult for GSD. Existing RCM methods do not take into
consideration these issues faced in GSD. Considering the state-of-the-art
in RCM, design and analysis of architecture, and cloud accountability,
this work contributes:
1. an alternative and novel mechanism for effective information and
knowledge-sharing towards RCM and traceability.
2. a novel methodology for the design and analysis of small-to-medium
size cloud-based systems, with a particular focus on the trade-off of
quality attributes.
3. a dependable framework that facilitates the RCM and traceability
method for cloud-based system engineering.
4. a novel methodology for assuring cloud accountability in terms of
dependability.
5. a cloud-based framework to facilitate the cloud accountability methodology.
The results show a traceable RCM linkage between system engineering
processes and stakeholder requirements for cloud-based GSD projects,
which is better than existing approaches. Also, the results show an improved dependability assurance of systems interfacing with the unpredictable cloud environment. We reach the conclusion that RCM with
a clear focus on traceability, which is then facilitated by a dependable
framework, improves the chance of developing a cloud-based GSD project
successfully
Self-adaptive Grid Resource Monitoring and discovery
The Grid provides a novel platform where the scientific and engineering communities can share data and computation across multiple administrative domains. There are several key services that must be offered by Grid middleware; one of them being the Grid Information Service( GIS). A GIS is a Grid middleware component which maintains information about hardware, software, services and people participating in a virtual organisation( VO). There is an inherent need in these systems for the delivery of reliable performance. This thesis describes a number of approaches which detail the development and application of a suite of benchmarks for the prediction of the process of resource discovery and monitoring on the Grid. A series of experimental studies of the characterisation of performance using benchmarking, are carried out. Several novel predictive algorithms are presented and evaluated in terms of their predictive error. Furthermore, predictive methods are developed which describe the behaviour of MDS2 for a variable number of user requests. The MDS is also extended to include job information from a local scheduler; this information is queried using requests of greatly varying complexity. The response of the MDS to these queries is then assessed in terms of several performance metrics.
The benchmarking of the dynamic nature of information within MDS3 which is based on the Open Grid Services Architecture (OGSA), and also the successor to MDS2, is also carried out. The performance of both the pull and push query mechanisms is analysed. GridAdapt (Self-adaptive Grid Resource Monitoring) is a new system that is proposed, built upon the Globus MDS3 benchmarking. It offers self-adaptation, autonomy and admission control at the Index Service, whilst ensuring that the MIDS is not overloaded and can meet its quality-of-service,f or example,i n terms of its average response time for servicing synchronous queries and the total number of queries returned per unit time
Air Force Institute of Technology Research Report 2007
This report summarizes the research activities of the Air Force Institute of Technology’s Graduate School of Engineering and Management. It describes research interests and faculty expertise; lists student theses/dissertations; identifies research sponsors and contributions; and outlines the procedures for contacting the school. Included in the report are: faculty publications, conference presentations, consultations, and funded research projects. Research was conducted in the areas of Aeronautical and Astronautical Engineering, Electrical Engineering and Electro-Optics, Computer Engineering and Computer Science, Systems and Engineering Management, Operational Sciences, Mathematics, Statistics and Engineering Physics
Advances in component-oriented programming
WCOP 2006 is the eleventh event in a series of highly successful
workshops, which took place in conjunction with every ECOOP
since 1996. Component oriented programming (COP) has been
described as the natural extension of object-oriented
programming to the realm of independently extensible
systems. Several important approaches have emerged over the
recent years, including component technology standards, such as
CORBA/CCM, COM/COM+, J2EE/EJB, and .NET, but also the increasing
appreciation of software architecture for component-based
systems, and the consequent effects on organizational processes
and structures as well as the software development business as a
whole.
COP aims at producing software components for a component market
and for late composition. Composers are third parties, possibly
the end users, who are not able or willing to change components.
This requires standards to allow independently created
components to interoperate, and specifications that put the
composer into the position to decide what can be composed under
which conditions. On these grounds, WCOP\u2796 led to the following
definition: "A component is a unit of composition with
contractually specified interfaces and explicit context
dependencies only. Components can be deployed independently and
are subject to composition by third parties."
After WCOP\u2796 focused on the fundamental terminology of COP, the
subsequent workshops expanded into the many related facets of
component software. WCOP 2006 emphasizes reasons for using
components beyond reuse. While considering software components
as a technical means to increase software reuse, other reasons
for investing into component technology tend to be overseen. For
example, components play an important role in frameworks and
product-lines to enable configurability (even if no component is
reused). Another role of components beyond reuse is to increase
the predictability of the properties of a system. The use of
components as contractually specified building blocks restricts
the degrees of freedom during software development compared to
classic line-by-line programming. This restriction is beneficial
for the predictability of system properties. For an engineering
approach to software design, it is important to understand the
implications of design decisions on a system\u27s properties.
Therefore, approaches to evaluate and predict properties of
systems by analyzing its components and its architecture are of
high interest.
To strengthen the relation between architectural descriptions of
systems and components, a comprehensible mapping to
component-oriented middleware platforms is important.
Model-driven development with its use of generators can
provide a suitable link between architectural views and
technical component execution platforms.
WCOP 2006 accepted 13 papers, which are organised according to
the program below. The organisers are looking forward to an
inspiring and thought provoking workshop. The organisers thank
Jens Happe and Michael Kuperberg for preparing
the proceedings volume
A semantic framework for event-driven service composition
Title from PDF of title page, viewed on September 14, 2011VitaDissertation advisor: Yugyung LeeIncludes bibliographical references (p. 289-329)Thesis (Ph.D)--School of Computing and Engineering. University of Missouri--Kansas City, 2011Service Oriented Architecture (SOA) has become a popular paradigm for designing
distributed systems where loosely coupled services (i.e. computational entities) can be
integrated seamlessly to provide complex composite services. Key challenges are discovery
of the required services using their formal descriptions and their coherent composition in a
timely manner. Most service descriptions are written in XML-based languages that are
syntactic, creating linguistic ambiguity during service matchmaking. Furthermore, existing
models that implement SOA have mostly middleware-controlled synchronous request/replybased
runtime binding of services that incur undesirable service latency. In addition, they
impose expensive state monitoring overhead on the middleware. Some newer event-driven
models introduce asynchronous publish/subscribe-based event notifications to consumer
applications and services. However, they require an event-library that stores definitions of
all possible system events, which is impractical in an open and dynamic system. The objective of this study is to efficiently address on-demand consumer requests
with minimum service latency and maximum consumer utility. It focuses on semantic eventdriven
service composition. For efficient semantic service discovery, the dissertation
proposes a novel service learning algorithm called Semantic Taxonomic Clustering (STC). The algorithm utilizes semantic service descriptions to cluster services into functional
categories for pruning search space during service discovery and composition. STC utilizes
a dynamic bit-encoding algorithm called DL-Encoding that enables linear time bit operationbased
semantic matchmaking as compared to expensive reasoner-based semantic
matchmaking. The algorithm shows significant improvement in performance and accuracy
over some of the important service category algorithms reported in the literature. A novel
user-friendly and computationally efficient query model called Desire-based Query
Model (DQM) is proposed for formally specifying service queries. STC and DQM serve as
the building block for the dual framework that is the core contribution of this dissertation: (i)
centralized ALNet (Activity Logic Network) platform and (ii) distributed agentbased
SMARTSPACE platform. The former incorporates a middleware controlled service
composition algorithm called ALNetComposer while the latter includes the SmartDeal
purely distributed composition algorithm. The query response accuracy and performance
were evaluated for both the algorithms under simulated event-driven SOA environments.
The experimental results show that various environmental parameters, such as domain
diversity and scope, size and complexity of the SOA system, and dynamicity of the SOA system, significantly affect accuracy and performance of the proposed model. This
dissertation demonstrates that the functionality and scalability of the proposed framework
are acceptable for relatively static and domain specific environments as well as large,
diverse, and highly dynamic environments. In summary, this dissertation addresses the key
design issues and problems in the area of asynchronous and pro-active event-driven service
composition.Introduction -- Research background -- Semantic service matchmaking & query modeling -- Service organization by learning service category -- ALNet: event-driven platform for service composition -- SMARTSPACE: distributed multi-agent based event-handeling -- Conclusion & future wor
A specification method for the scalable self-governance of complex autonomic systems
IBM, amongst many others, have sought to endow computer systems with selfmanagement capabilities by delegating vital functions to the software itself and proposed the Autonomic Computing model. Hence inducing the so-called self-* properties including the system's ability to be self-configuring, self-optimising, self-healing and self-protecting. Initial attempts to realise such a vision have so far mostly relied on a passive adaptation whereby Design by Contract and Event-Condition-Action (ECA) type constructs are used to regulate the target systems behaviour: When a specific event makes a certain condition true then an action is triggered which executes either within the system or on its environment Whilst, such a model works well for closed systems, its effectiveness and applicability of approach diminishes as the size and complexity of the managed system increases, necessitating frequent updates to the ECA rule set to cater for new and/or unforeseen systems' behaviour. More recent research works are now adopting the parametric adaptation model, where the events, conditions and actions may be adjusted at runtime in response to the system's observed state. Such an improved control model works well up to a point, but for large scale systems of systems, with very many component interactions, the predictability and traceability of the regulation and its impact on the whole system is intractable. The selforganising systems theory, however, offers a scaleable alternative to systems control utilising emerging behaviour, observed at a global level, resulting from the low-level interactions of the distributed components. Whereby, for instance, key signals (signs) for ECA style feedback control need no longer be recognised or understood in the context of the design time system but are defined by their relevance to the runtime system. Nonetheless this model still suffers from a usually inaccessible control model with no intrinsic meaning assigned to data extraction from the systems operation. In other words, there is no grounded definition of particular observable events occurring in the system. This condition is termed the Signal Grounding Problem. This problem cannot usually be solved by analytical or algorithmic methods, as these solutions generally require precise problem formulations and a static operating domain. Rather cognitive techniques will be needed that perform effectively to evaluate and improve performance in the presence of complex, incomplete, dynamic and evolving environments. In order to develop a specification method for scalable self-governance of autonomic systems of systems, this thesis presents a number of ways to alleviate, or circumvent, the Signal Grounding Problem through the utilisation of cognitive systems and the properties of complex systems. After reviewing the specification methods available for governance models, the Situation Calculus dialect of first order logic is described with the necessary modalities for the specification of deliberative monitoring in partially observable environments with stochastic actions. This permits a specification method that allows the depiction of system guards and norms, under central control, as well as the deliberative functions required for decentralised components to present techniques around the Signal Grounding problem, engineer emergence and generally utilise the properties of large complex systems for their own self-governance. It is shown how these large-scale behaviours may be implemented and the properties assessed and utilised by an Observer System through fully functioning implementations and simulations. The work concludes with two case studies showing how the specification would be achieved in practice: An observer based meta-system for a decision support system in medicine is described, specified and implemented up to parametric adaptation and a NASA project is described with a specification given for the interactions and cooperative behaviour that leads to scale-free connectivity, which the observer system may then utilise for a previously described efficient monitoring strategy
- …