136 research outputs found

    COLEG: Collaborative Learning Environment within Grid

    Get PDF
    The principal function of the CSCL environments is to provide to the various users (students, teachers, tutors…), the best activities with the best tools at the best time according to their needs. If a CSCL system is a collection of activities or learning process, we can cut out its functionalities in a certain number of autonomous functions which can then be carried out separately in the form of autonomous applications by using the technology of the Web/Grid services. The emerging technologies based on the Grid are increasingly being adopted to improve education and provide better services for learning. These services are offered to students who, regardless of their computer systems, can collaborate to improve their cognitive and social skills. This article presents COLEG (COllaborative Learning Environment within Grid), which aims to employ the capacities offered by the Grid to give the various actors, all the power of learning, collaboration and communication in an adaptable, heterogeneous and dynamic sight

    Proceedings of the 2005 IJCAI Workshop on AI and Autonomic Communications

    Get PDF

    Interim research assessment 2003-2005 - Computer Science

    Get PDF
    This report primarily serves as a source of information for the 2007 Interim Research Assessment Committee for Computer Science at the three technical universities in the Netherlands. The report also provides information for others interested in our research activities

    A reactive architecture for cloud-based system engineering

    Get PDF
    PhD ThesisSoftware system engineering is increasingly practised over globally distributed locations. Such a practise is termed as Global Software Development (GSD). GSD has become a business necessity mainly because of the scarcity of resources, cost, and the need to locate development closer to the customers. GSD is highly dependent on requirements management, but system requirements continuously change. Poorly managed change in requirements affects the overall cost, schedule and quality of GSD projects. It is particularly challenging to manage and trace such changes, and hence we require a rigorous requirement change management (RCM) process. RCM is not trivial in collocated software development; and with the presence of geographical, cultural, social and temporal factors, it makes RCM profoundly difficult for GSD. Existing RCM methods do not take into consideration these issues faced in GSD. Considering the state-of-the-art in RCM, design and analysis of architecture, and cloud accountability, this work contributes: 1. an alternative and novel mechanism for effective information and knowledge-sharing towards RCM and traceability. 2. a novel methodology for the design and analysis of small-to-medium size cloud-based systems, with a particular focus on the trade-off of quality attributes. 3. a dependable framework that facilitates the RCM and traceability method for cloud-based system engineering. 4. a novel methodology for assuring cloud accountability in terms of dependability. 5. a cloud-based framework to facilitate the cloud accountability methodology. The results show a traceable RCM linkage between system engineering processes and stakeholder requirements for cloud-based GSD projects, which is better than existing approaches. Also, the results show an improved dependability assurance of systems interfacing with the unpredictable cloud environment. We reach the conclusion that RCM with a clear focus on traceability, which is then facilitated by a dependable framework, improves the chance of developing a cloud-based GSD project successfully

    Self-adaptive Grid Resource Monitoring and discovery

    Get PDF
    The Grid provides a novel platform where the scientific and engineering communities can share data and computation across multiple administrative domains. There are several key services that must be offered by Grid middleware; one of them being the Grid Information Service( GIS). A GIS is a Grid middleware component which maintains information about hardware, software, services and people participating in a virtual organisation( VO). There is an inherent need in these systems for the delivery of reliable performance. This thesis describes a number of approaches which detail the development and application of a suite of benchmarks for the prediction of the process of resource discovery and monitoring on the Grid. A series of experimental studies of the characterisation of performance using benchmarking, are carried out. Several novel predictive algorithms are presented and evaluated in terms of their predictive error. Furthermore, predictive methods are developed which describe the behaviour of MDS2 for a variable number of user requests. The MDS is also extended to include job information from a local scheduler; this information is queried using requests of greatly varying complexity. The response of the MDS to these queries is then assessed in terms of several performance metrics. The benchmarking of the dynamic nature of information within MDS3 which is based on the Open Grid Services Architecture (OGSA), and also the successor to MDS2, is also carried out. The performance of both the pull and push query mechanisms is analysed. GridAdapt (Self-adaptive Grid Resource Monitoring) is a new system that is proposed, built upon the Globus MDS3 benchmarking. It offers self-adaptation, autonomy and admission control at the Index Service, whilst ensuring that the MIDS is not overloaded and can meet its quality-of-service,f or example,i n terms of its average response time for servicing synchronous queries and the total number of queries returned per unit time

    Air Force Institute of Technology Research Report 2007

    Get PDF
    This report summarizes the research activities of the Air Force Institute of Technology’s Graduate School of Engineering and Management. It describes research interests and faculty expertise; lists student theses/dissertations; identifies research sponsors and contributions; and outlines the procedures for contacting the school. Included in the report are: faculty publications, conference presentations, consultations, and funded research projects. Research was conducted in the areas of Aeronautical and Astronautical Engineering, Electrical Engineering and Electro-Optics, Computer Engineering and Computer Science, Systems and Engineering Management, Operational Sciences, Mathematics, Statistics and Engineering Physics

    Advances in component-oriented programming

    Get PDF
    WCOP 2006 is the eleventh event in a series of highly successful workshops, which took place in conjunction with every ECOOP since 1996. Component oriented programming (COP) has been described as the natural extension of object-oriented programming to the realm of independently extensible systems. Several important approaches have emerged over the recent years, including component technology standards, such as CORBA/CCM, COM/COM+, J2EE/EJB, and .NET, but also the increasing appreciation of software architecture for component-based systems, and the consequent effects on organizational processes and structures as well as the software development business as a whole. COP aims at producing software components for a component market and for late composition. Composers are third parties, possibly the end users, who are not able or willing to change components. This requires standards to allow independently created components to interoperate, and specifications that put the composer into the position to decide what can be composed under which conditions. On these grounds, WCOP\u2796 led to the following definition: "A component is a unit of composition with contractually specified interfaces and explicit context dependencies only. Components can be deployed independently and are subject to composition by third parties." After WCOP\u2796 focused on the fundamental terminology of COP, the subsequent workshops expanded into the many related facets of component software. WCOP 2006 emphasizes reasons for using components beyond reuse. While considering software components as a technical means to increase software reuse, other reasons for investing into component technology tend to be overseen. For example, components play an important role in frameworks and product-lines to enable configurability (even if no component is reused). Another role of components beyond reuse is to increase the predictability of the properties of a system. The use of components as contractually specified building blocks restricts the degrees of freedom during software development compared to classic line-by-line programming. This restriction is beneficial for the predictability of system properties. For an engineering approach to software design, it is important to understand the implications of design decisions on a system\u27s properties. Therefore, approaches to evaluate and predict properties of systems by analyzing its components and its architecture are of high interest. To strengthen the relation between architectural descriptions of systems and components, a comprehensible mapping to component-oriented middleware platforms is important. Model-driven development with its use of generators can provide a suitable link between architectural views and technical component execution platforms. WCOP 2006 accepted 13 papers, which are organised according to the program below. The organisers are looking forward to an inspiring and thought provoking workshop. The organisers thank Jens Happe and Michael Kuperberg for preparing the proceedings volume

    A semantic framework for event-driven service composition

    Get PDF
    Title from PDF of title page, viewed on September 14, 2011VitaDissertation advisor: Yugyung LeeIncludes bibliographical references (p. 289-329)Thesis (Ph.D)--School of Computing and Engineering. University of Missouri--Kansas City, 2011Service Oriented Architecture (SOA) has become a popular paradigm for designing distributed systems where loosely coupled services (i.e. computational entities) can be integrated seamlessly to provide complex composite services. Key challenges are discovery of the required services using their formal descriptions and their coherent composition in a timely manner. Most service descriptions are written in XML-based languages that are syntactic, creating linguistic ambiguity during service matchmaking. Furthermore, existing models that implement SOA have mostly middleware-controlled synchronous request/replybased runtime binding of services that incur undesirable service latency. In addition, they impose expensive state monitoring overhead on the middleware. Some newer event-driven models introduce asynchronous publish/subscribe-based event notifications to consumer applications and services. However, they require an event-library that stores definitions of all possible system events, which is impractical in an open and dynamic system. The objective of this study is to efficiently address on-demand consumer requests with minimum service latency and maximum consumer utility. It focuses on semantic eventdriven service composition. For efficient semantic service discovery, the dissertation proposes a novel service learning algorithm called Semantic Taxonomic Clustering (STC). The algorithm utilizes semantic service descriptions to cluster services into functional categories for pruning search space during service discovery and composition. STC utilizes a dynamic bit-encoding algorithm called DL-Encoding that enables linear time bit operationbased semantic matchmaking as compared to expensive reasoner-based semantic matchmaking. The algorithm shows significant improvement in performance and accuracy over some of the important service category algorithms reported in the literature. A novel user-friendly and computationally efficient query model called Desire-based Query Model (DQM) is proposed for formally specifying service queries. STC and DQM serve as the building block for the dual framework that is the core contribution of this dissertation: (i) centralized ALNet (Activity Logic Network) platform and (ii) distributed agentbased SMARTSPACE platform. The former incorporates a middleware controlled service composition algorithm called ALNetComposer while the latter includes the SmartDeal purely distributed composition algorithm. The query response accuracy and performance were evaluated for both the algorithms under simulated event-driven SOA environments. The experimental results show that various environmental parameters, such as domain diversity and scope, size and complexity of the SOA system, and dynamicity of the SOA system, significantly affect accuracy and performance of the proposed model. This dissertation demonstrates that the functionality and scalability of the proposed framework are acceptable for relatively static and domain specific environments as well as large, diverse, and highly dynamic environments. In summary, this dissertation addresses the key design issues and problems in the area of asynchronous and pro-active event-driven service composition.Introduction -- Research background -- Semantic service matchmaking & query modeling -- Service organization by learning service category -- ALNet: event-driven platform for service composition -- SMARTSPACE: distributed multi-agent based event-handeling -- Conclusion & future wor

    A specification method for the scalable self-governance of complex autonomic systems

    Get PDF
    IBM, amongst many others, have sought to endow computer systems with selfmanagement capabilities by delegating vital functions to the software itself and proposed the Autonomic Computing model. Hence inducing the so-called self-* properties including the system's ability to be self-configuring, self-optimising, self-healing and self-protecting. Initial attempts to realise such a vision have so far mostly relied on a passive adaptation whereby Design by Contract and Event-Condition-Action (ECA) type constructs are used to regulate the target systems behaviour: When a specific event makes a certain condition true then an action is triggered which executes either within the system or on its environment Whilst, such a model works well for closed systems, its effectiveness and applicability of approach diminishes as the size and complexity of the managed system increases, necessitating frequent updates to the ECA rule set to cater for new and/or unforeseen systems' behaviour. More recent research works are now adopting the parametric adaptation model, where the events, conditions and actions may be adjusted at runtime in response to the system's observed state. Such an improved control model works well up to a point, but for large scale systems of systems, with very many component interactions, the predictability and traceability of the regulation and its impact on the whole system is intractable. The selforganising systems theory, however, offers a scaleable alternative to systems control utilising emerging behaviour, observed at a global level, resulting from the low-level interactions of the distributed components. Whereby, for instance, key signals (signs) for ECA style feedback control need no longer be recognised or understood in the context of the design time system but are defined by their relevance to the runtime system. Nonetheless this model still suffers from a usually inaccessible control model with no intrinsic meaning assigned to data extraction from the systems operation. In other words, there is no grounded definition of particular observable events occurring in the system. This condition is termed the Signal Grounding Problem. This problem cannot usually be solved by analytical or algorithmic methods, as these solutions generally require precise problem formulations and a static operating domain. Rather cognitive techniques will be needed that perform effectively to evaluate and improve performance in the presence of complex, incomplete, dynamic and evolving environments. In order to develop a specification method for scalable self-governance of autonomic systems of systems, this thesis presents a number of ways to alleviate, or circumvent, the Signal Grounding Problem through the utilisation of cognitive systems and the properties of complex systems. After reviewing the specification methods available for governance models, the Situation Calculus dialect of first order logic is described with the necessary modalities for the specification of deliberative monitoring in partially observable environments with stochastic actions. This permits a specification method that allows the depiction of system guards and norms, under central control, as well as the deliberative functions required for decentralised components to present techniques around the Signal Grounding problem, engineer emergence and generally utilise the properties of large complex systems for their own self-governance. It is shown how these large-scale behaviours may be implemented and the properties assessed and utilised by an Observer System through fully functioning implementations and simulations. The work concludes with two case studies showing how the specification would be achieved in practice: An observer based meta-system for a decision support system in medicine is described, specified and implemented up to parametric adaptation and a NASA project is described with a specification given for the interactions and cooperative behaviour that leads to scale-free connectivity, which the observer system may then utilise for a previously described efficient monitoring strategy
    • …
    corecore