15 research outputs found

    Reactivity on the Web

    Get PDF
    Reactivity, the ability to detect simple and composite events and respond in a timely manner, is an essential requirement in many present-day information systems. With the emergence of new, dynamic Web applications, reactivity on the Web is receiving increasing attention. Reactive Web-based systems need to detect and react not only to simple events but also to complex, real-life situations. This paper introduces XChange, a language for programming reactive behaviour on the Web, emphasising the querying of event data and detection of composite events

    State-of-the-art on evolution and reactivity

    Get PDF
    This report starts by, in Chapter 1, outlining aspects of querying and updating resources on the Web and on the Semantic Web, including the development of query and update languages to be carried out within the Rewerse project. From this outline, it becomes clear that several existing research areas and topics are of interest for this work in Rewerse. In the remainder of this report we further present state of the art surveys in a selection of such areas and topics. More precisely: in Chapter 2 we give an overview of logics for reasoning about state change and updates; Chapter 3 is devoted to briefly describing existing update languages for the Web, and also for updating logic programs; in Chapter 4 event-condition-action rules, both in the context of active database systems and in the context of semistructured data, are surveyed; in Chapter 5 we give an overview of some relevant rule-based agents frameworks

    State-of-the-art on evolution and reactivity

    Get PDF
    This report starts by, in Chapter 1, outlining aspects of querying and updating resources on the Web and on the Semantic Web, including the development of query and update languages to be carried out within the Rewerse project. From this outline, it becomes clear that several existing research areas and topics are of interest for this work in Rewerse. In the remainder of this report we further present state of the art surveys in a selection of such areas and topics. More precisely: in Chapter 2 we give an overview of logics for reasoning about state change and updates; Chapter 3 is devoted to briefly describing existing update languages for the Web, and also for updating logic programs; in Chapter 4 event-condition-action rules, both in the context of active database systems and in the context of semistructured data, are surveyed; in Chapter 5 we give an overview of some relevant rule-based agents frameworks

    Managing complex taxonomic data in an object-oriented database.

    Get PDF
    This thesis addresses the problem of multiple overlapping classifications in object-oriented databases through the example of plant taxonomy. These multiple overlapping classifications are independent simple classifications that share information (nodes and leaves), therefore overlap. Plant taxonomy was chosen as the motivational application domain because taxonomic classifications are especially complex and have changed over long periods of time, therefore overlap in a significant manner. This work extracts basic requirements for the support of multiple overlapping classifications in general, and in the context of plant taxonomy in particular. These requirements form the basis on which a prototype is defmed and built. The prototype, an extended object-oriented database, is extended from an object-oriented model based on ODMG through the provision of a relationship management mechanism. These relationships form the main feature used to build classifications. This emphasis on relationships allows the description of classifications orthogonal to the classified data (for reuse and integration of the mechanism with existing databases and for classification of non co-operating data), and allows an easier and more powerful management of semantic data (both within and without a classification). Additional mechanisms such as integrity constraints are investigated and implemented. Finally, the implementation of the prototype is presented and is evaluated, from the point of view of both usability and expressiveness (using plant taxonomy as an application), and its performance as a database system. This evaluation shows that the prototype meets the needs of taxonomists

    A trusted infrastructure for symbolic analysis of event-based web APIs

    Get PDF
    JavaScript has been widely adopted for the development of Web applications, being used for both client and server-side code. Client-side JavaScript programs commonly interact with Web APIs, for instance, to capture the user interaction with the Web page via events. The use of such APIs increases the complexity of JavaScript programs. In fact, most errors in these programs are caused by the misuse of Web APIs. There are several approaches for detecting errors in client-side JavaScript programs, but they either assume the use of a single API or do not model APIs faithfully, giving rise to inconsistent behaviour and lack of trust. We address the problem by developing a trustworthy infrastructure for the static analysis of Web APIs. We focus on two aspects of JavaScript programs: event-driven and message-passing programming, as these paradigms are common sources of confusion among developers. We choose to target the DOM event model and the JavaScript Promises and JavaScript async/await, which facilitate event-driven programming. Additionally, we target the message-passing model of the WebMessaging and WebWorkers APIs. We design formal semantics for events and message-passing to capture fundamental operations required by those APIs, and API reference implementations which are trustworthy in that they follow the respective standards and have been thoroughly tested against their official test suites. Using our formal semantics and reference implementations, we develop JaVerT.Click, the first static symbolic execution tool for JavaScript supporting both event-based and message-passing APIs. We evaluated both the reference implementations and the symbolic execution engine of JaVerT.Click. By testing the reference implementations against their official test suites, we found coverage gaps and issues in the test suites, most of which have been since fixed. By testing the symbolic execution engine against three open-source libraries, we established the bounded correctness of functional properties and found real bugs.Open Acces

    Personnalisation d'analyses décisionnelles sur des données multidimensionnelles

    Get PDF
    This thesis investigates OLAP analysis personalization within multidimensional databases. OLAP analyse is modeled through a graph where nodes represent the analysis contexts and graph edges represent the user operations. The analysis context regroups the user query as well as result. It is well described by a specific tree structure that is independent on the visualization structures of data and query languages. We provided a model for user preferences on the multidimensional schema and values. Each preference is associated with a specific analysis context. Based on previous models, we proposed a generic framework that includes two personalization processes. First process, denoted query personalization, aims to enhancing user query with related preferences in order to produce a new one that generates a personalized result. Second personalization process is query recommendation that allows helping user throughout the OLAP data exploration phase. Our recommendation framework supports three recommendation scenarios, i.e., assisting user in query composition, suggesting the forthcoming query, and suggesting alternative queries. Recommendations are built progressively basing on user preferences. In order to implement our framework, we developed a prototype system that supports query personalization and query recommendation processes. We present experimental results showing the efficiency and the effectiveness of our approaches.Le travail prĂ©sentĂ© dans cette thĂšse aborde la problĂ©matique de la personnalisation des analyses OLAP au sein des bases de donnĂ©es multidimensionnelles. Une analyse OLAP est modĂ©lisĂ©e par un graphe dont les noeuds reprĂ©sentent les contextes d'analyse et les arcs traduisent les opĂ©rations de l'utilisateur. Le contexte d'analyse regroupe la requĂȘte et le rĂ©sultat. Il est dĂ©crit par un arbre spĂ©cifique qui est indĂ©pendant des structures de visualisation des donnĂ©es et des langages de requĂȘte. Par ailleurs, nous proposons un modĂšle de prĂ©fĂ©rences utilisateur exprimĂ©es sur le schĂ©ma multidimensionnel et sur les valeurs. Chaque prĂ©fĂ©rence est associĂ©e Ă  un contexte d'analyse particulier. En nous basant sur ces modĂšles, nous proposons un cadre gĂ©nĂ©rique comportant deux mĂ©canismes de personnalisation. Le premier mĂ©canisme est la personnalisation de requĂȘte. Il permet d'enrichir la requĂȘte utilisateur Ă  l'aide des prĂ©fĂ©rences correspondantes afin de gĂ©nĂ©rer un rĂ©sultat qui satisfait au mieux aux besoins de l'usager. Le deuxiĂšme mĂ©canisme de personnalisation est la recommandation de requĂȘtes qui permet d'assister l'utilisateur tout au long de son exploration des donnĂ©es OLAP. Trois scĂ©narios de recommandation sont dĂ©finis : l'assistance Ă  la formulation de requĂȘte, la proposition de la prochaine requĂȘte et la suggestion de requĂȘtes alternatives. Ces recommandations sont construites progressivement Ă  l'aide des prĂ©fĂ©rences de l'utilisateur. Afin valider nos diffĂ©rentes contributions, nous avons dĂ©veloppĂ© un prototype qui intĂšgre les mĂ©canismes de personnalisation et de recommandation de requĂȘte proposĂ©s. Nous prĂ©sentons les rĂ©sultats d'expĂ©rimentations montrant la performance et l'efficacitĂ© de nos approches. Mots-clĂ©s: OLAP, analyse dĂ©cisionnelle, personnalisation de requĂȘte, systĂšme de recommandation, prĂ©fĂ©rence utilisateur, contexte d'analyse, appariement d'arbres de contexte

    Model driven design and data integration in semantic web information systems

    Get PDF
    The Web is quickly evolving in many ways. It has evolved from a Web of documents into a Web of applications in which a growing number of designers offer new and interactive Web applications with people all over the world. However, application design and implementation remain complex, error-prone and laborious. In parallel there is also an evolution from a Web of documents into a Web of `knowledge' as a growing number of data owners are sharing their data sources with a growing audience. This brings the potential new applications for these data sources, including scenarios in which these datasets are reused and integrated with other existing and new data sources. However, the heterogeneity of these data sources in syntax, semantics and structure represents a great challenge for application designers. The Semantic Web is a collection of standards and technologies that offer solutions for at least the syntactic and some structural issues. If offers semantic freedom and flexibility, but this leaves the issue of semantic interoperability. In this thesis we present Hera-S, an evolution of the Model Driven Web Engineering (MDWE) method Hera. MDWEs allow designers to create data centric applications using models instead of programming. Hera-S especially targets Semantic Web sources and provides a flexible method for designing personalized adaptive Web applications. Hera-S defines several models that together define the target Web application. Moreover we implemented a framework called Hydragen, which is able to execute the Hera-S models to run the desired Web application. Hera-S' core is the Application Model (AM) in which the main logic of the application is defined, i.e. defining the groups of data elements that form logical units or subunits, the personalization conditions, and the relationships between the units. Hera-S also uses a so-called Domain Model (DM) that describes the content and its structure. However, this DM is not Hera-S specific, but instead allows any Semantic Web source representation as its DM, as long as its content can be queried by the standardized Semantic Web query language SPARQL. The same holds for the User Model (UM). The UM can be used for personalization conditions, but also as a source of user-related content if necessary. In fact, the difference between DM and UM is conceptual as their implementation within Hydragen is the same. Hera-S also defines a presentation model (PM) which defines presentation details of elements like order and style. In order to help designers with building their Web applications we have introduced a toolset, Hera Studio, which allows to build the different models graphically. Hera Studio also provides some additional functionality like model checking and deployment of the models in Hydragen. Both Hera-S and its implementation Hydragen are designed to be flexible regarding the user of models. In order to achieve this Hydragen is a stateless engine that queries for relevant information from the models at every page request. This allows the models and data to be changed in the datastore during runtime. We show that one way to exploit this flexibility is by applying aspect-orientation to the AM. Aspect-orientation allows us to dynamically inject functionality that pervades the entire application. Another way to exploit Hera-S' flexibility is in reusing specialized components, e.g. for presentation generation. We present a configuration of Hydragen in which we replace our native presentation generation functionality by the AMACONT engine. AMACONT provides more extensive multi-level presentation generation and adaptation capabilities as well aspect-orientation and a form of semantic based adaptation. Hera-S was designed to allow the (re-)use of any (Semantic) Web datasource. It even opens up the possibility for data integration at the back end, by using an extendible storage layer in our database of choice Sesame. However, even though theoretically possible it still leaves much of the actual data integration issue. As this is a recurring issue in many domains, a broader challenge than for Hera-S design only, we decided to look at this issue in isolation. We present a framework called Relco which provides a language to express data transformation operations as well as a collection of techniques that can be used to (semi-)automatically find relationships between concepts in different ontologies. This is done with a combination of syntactic, semantic and collaboration techniques, which together provide strong clues for which concepts are most likely related. In order to prove the applicability of Relco we explore five application scenarios in different domains for which data integration is a central aspect. This includes a cultural heritage portal, Explorer, for which data from several datasources was integrated and was made available by a mapview, a timeline and a graph view. Explorer also allows users to provide metadata for objects via a tagging mechanism. Another application is SenSee: an electronic TV-guide and recommender. TV-guide data was integrated and enriched with semantically structured data from several sources. Recommendations are computed by exploiting the underlying semantic structure. ViTa was a project in which several techniques for tagging and searching educational videos were evaluated. This includes scenarios in which user tags are related with an ontology, or other tags, using the Relco framework. The MobiLife project targeted the facilitation of a new generation of mobile applications that would use context-based personalization. This can be done using a context-based user profiling platform that can also be used for user model data exchange between mobile applications using technologies like Relco. The final application scenario that is shown is from the GRAPPLE project which targeted the integration of adaptive technology into current learning management systems. A large part of this integration is achieved by using a user modeling component framework in which any application can store user model information, but which can also be used for the exchange of user model data

    Actes des 2Ăšmes journĂ©es sur l’IngĂ©nierie DirigĂ©e

    Get PDF
    National audienceL’ingĂ©nierie dirigĂ©e par les modĂšles (IDM), appelĂ©e en anglais MDE (Model-Driven Engineering) ou aussi MDD (Model-Driven Development) place le modĂšle au centre du processus de conception et permet Ă  cette notion de modĂšle de passer d’un rĂŽle contemplatif Ă  un rĂŽle unificateur vis-Ă -vis des autres activitĂ©s du cycle de dĂ©veloppement du logiciel. L’IDM doit alors ĂȘtre vu non pas comme une rĂ©volution, mais comme un moyen d’intĂ©grationde diffĂ©rents espaces techniques pour aller vers une production automatisĂ©e des logiciels.L’ingĂ©nierie dirigĂ©e par les modĂšles apporte alors des solutions Ă  la construction de ces nouveaux logiciels en proposant des approches de modĂ©lisation, de mĂ©tamodĂ©lisation, de dĂ©termination du domaine, de transformation et de prise en compte des plates-formes. Ces approches sont accompagnĂ©es de dĂ©marches de conception et de moyens de gĂ©nĂ©ration de code, mais Ă©galement de validation et de vĂ©rification de la conformitĂ© des modĂšles produits vis-Ă -vis des mĂ©tamodĂšles. Elles sont proches des idĂ©es actuelles comme la programmation gĂ©nĂ©rative, les langages spĂ©cifiques de domaine (DSL), le MIC (Model Integrating Computing) ou encore les usines Ă  logiciels (Software factories). AprĂšs le succĂšs des journĂ©es IDM Ă  Paris en 2005, la seconde Ă©dition de ces journĂ©es se dĂ©roule Ă  Lille et a pour objectif de rassembler les chercheurs francophones intĂ©ressĂ©s par ce domaine et souhaitant participer Ă  la structuration de cette communautĂ© scientifique Ă©mergente
    corecore