8 research outputs found

    Seventh Biennial Report : June 2003 - March 2005

    No full text

    Sixth Biennial Report : August 2001 - May 2003

    No full text

    Eight Biennial Report : April 2005 – March 2007

    No full text

    Advanced methods for query routing in peer-to-peer information retrieval

    Get PDF
    One of the most challenging problems in peer-to-peer networks is query routing: effectively and efficiently identifying peers that can return high-quality local results for a given query. Existing methods from the areas of distributed information retrieval and metasearch engines do not adequately address the peculiarities of a peer-to-peer network. The main contributions of this thesis are as follows: 1. Methods for query routing that take into account the mutual overlap of different peers\u27; collections, 2. Methods for query routing that take into account the correlations between multiple terms, 3. Comparative evaluation of different query routing methods. Our experiments confirm the superiority of our novel query routing methods over the prior state-of-the-art, in particular in the context of peer-to-peer Web search.Eines der drĂ€ngendsten Probleme in Peer-to-Peer-Netzwerken ist Query-Routing: das effektive und effiziente Identifizieren solcher Peers, die qualitativ hochwertige lokale Ergebnisse zu einer gegebenen Anfrage liefern können. Die bisher bekannten Verfahren aus dem Bereich der verteilten Informationssuche sowie der Metasuchmaschinen werden den Besonderheiten von Peer-to-Peer-Netzwerken nicht gerecht. Die HautbeitrĂ€ge dieser Arbeit teilen sich in folgende Schwerpunkte: 1. Query-Routing unter BerĂŒcksichtigung der gegenseitigen ĂŒberlappung der Kollektionen verschiedener Peers, 2. Query-Routing unter BerĂŒcksichtigung der Korrelationen zwischen verschiedenen Termen, 3. Vergleichende Evaluierung verschiedener Methoden zum Query-Routing. Unsere Experimente bestĂ€tigen die Überlegenheit der in dieser Arbeit entwickelten Verfahren gegenĂŒber den bisher bekannten Verfahren, insbesondere im Kontext von Peer-to-Peer-Websuche

    Äriprotsessi tulemuste ennustav ja korralduslik seire

    Get PDF
    Viimastel aastatel on erinevates valdkondades tegutsevad ettevĂ”tted ĂŒles nĂ€idanud kasvavat huvi masinĂ”ppel pĂ”hinevate rakenduste kasutusele vĂ”tmiseks. Muuhulgas otsitakse vĂ”imalusi oma Ă€riprotsesside efektiivsuse tĂ”stmiseks, kasutades ennustusmudeleid protsesside jooksvaks seireks. Sellised ennustava protsessiseire meetodid vĂ”tavad sisendiks sĂŒndmuslogi, mis koosneb hulgast lĂ”petatud Ă€riprotsessi juhtumite sĂŒndmusjadadest, ning kasutavad masinĂ”ppe algoritme ennustusmudelite treenimiseks. Saadud mudelid teevad ennustusi lĂ”petamata (antud ajahetkel aktiivsete) protsessijuhtumite jaoks, vĂ”ttes sisendiks sĂŒndmuste jada, mis selle hetkeni on toimunud ning ennustades kas jĂ€rgmist sĂŒndmust antud juhtumis, juhtumi lĂ”ppemiseni jÀÀnud aega vĂ”i instantsi lĂ”pptulemust. LĂ”pptulemusele orienteeritud ennustava protsessiseire meetodid keskenduvad ennustamisele, kas protsessijuhtum lĂ”ppeb soovitud vĂ”i ebasoovitava lĂ”pptulemusega. SĂŒsteemi kasutaja saab ennustuste alusel otsustada, kas sekkuda antud protsessijuhtumisse vĂ”i mitte, eesmĂ€rgiga Ă€ra hoida ebasoovitavat lĂ”pptulemust vĂ”i leevendada selle negatiivseid tagajĂ€rgi. Erinevalt puhtalt ennustavatest sĂŒsteemidest annavad korralduslikud protsessiseire meetodid kasutajale ka soovitusi, kas ja kuidas antud juhtumisse sekkuda, eesmĂ€rgiga optimeerida mingit kindlat kasulikkusfunktsiooni. KĂ€esolev doktoritöö uurib, kuidas treenida, hinnata ja kasutada ennustusmudeleid Ă€riprotsesside lĂ”pptulemuste ennustava ja korraldusliku seire raames. Doktoritöö pakub vĂ€lja taksonoomia olemasolevate meetodite klassifitseerimiseks ja vĂ”rdleb neid katseliselt. Lisaks pakub töö vĂ€lja raamistiku tekstiliste andmete kasutamiseks antud ennustusmudelites. Samuti pakume vĂ€lja ennustuste ajalise stabiilsuse mĂ”iste ning koostame raamistiku korralduslikuks protsessiseireks, mis annab kasutajatele soovitusi, kas protsessi sekkuda vĂ”i mitte. Katsed nĂ€itavad, et vĂ€ljapakutud lahendused tĂ€iendavad olemasolevaid meetodeid ning aitavad kaasa ennustava protsessiseire sĂŒsteemide rakendamisele reaalsetes sĂŒsteemides.Recent years have witnessed a growing adoption of machine learning techniques for business improvement across various fields. Among other emerging applications, organizations are exploiting opportunities to improve the performance of their business processes by using predictive models for runtime monitoring. Such predictive process monitoring techniques take an event log (a set of completed business process execution traces) as input and use machine learning techniques to train predictive models. At runtime, these techniques predict either the next event, the remaining time, or the final outcome of an ongoing case, given its incomplete execution trace consisting of the events performed up to the present moment in the given case. In particular, a family of techniques called outcome-oriented predictive process monitoring focuses on predicting whether a case will end with a desired or an undesired outcome. The user of the system can use the predictions to decide whether or not to intervene, with the purpose of preventing an undesired outcome or mitigating its negative effects. Prescriptive process monitoring systems go beyond purely predictive ones, by not only generating predictions but also advising the user if and how to intervene in a running case in order to optimize a given utility function. This thesis addresses the question of how to train, evaluate, and use predictive models for predictive and prescriptive monitoring of business process outcomes. The thesis proposes a taxonomy and performs a comparative experimental evaluation of existing techniques in the field. Moreover, we propose a framework for incorporating textual data to predictive monitoring systems. We introduce the notion of temporal stability to evaluate these systems and propose a prescriptive process monitoring framework for advising users if and how to act upon the predictions. The results suggest that the proposed solutions complement the existing techniques and can be useful for practitioners in implementing predictive process monitoring systems in real life

    Cost- and workload-driven data management in the cloud

    Get PDF
    This thesis deals with the challenge of finding the right balance between consistency, availability, latency and costs, captured by the CAP/PACELC trade-offs, in the context of distributed data management in the Cloud. At the core of this work, cost and workload-driven data management protocols, called CCQ protocols, are developed. First, this includes the development of C3, which is an adaptive consistency protocol that is able to adjust consistency at runtime by considering consistency and inconsistency costs. Second, the development of Cumulus, an adaptive data partitioning protocol, that can adapt partitions by considering the application workload so that expensive distributed transactions are minimized or avoided. And third, the development of QuAD, a quorum-based replication protocol, that constructs the quorums in such a way so that, given a set of constraints, the best possible performance is achieved. The behavior of each CCQ protocol is steered by a cost model, which aims at reducing the costs and overhead for providing the desired data management guarantees. The CCQ protocols are able to continuously assess their behavior, and if necessary to adapt the behavior at runtime based on application workload and the cost model. This property is crucial for applications deployed in the Cloud, as they are characterized by a highly dynamic workload, and high scalability and availability demands. The dynamic adaptation of the behavior at runtime does not come for free, and may generate considerable overhead that might outweigh the gain of adaptation. The CCQ cost models incorporate a control mechanism, which aims at avoiding expensive and unnecessary adaptations, which do not provide any benefits to applications. The adaptation is a distributed activity that requires coordination between the sites in a distributed database system. The CCQ protocols implement safe online adaptation approaches, which exploit the properties of 2PC and 2PL to ensure that all sites behave in accordance with the cost model, even in the presence of arbitrary failures. It is crucial to guarantee a globally consistent view of the behavior, as in contrary the effects of the cost models are nullified. The presented protocols are implemented as part of a prototypical database system. Their modular architecture allows for a seamless extension of the optimization capabilities at any level of their implementation. Finally, the protocols are quantitatively evaluated in a series of experiments executed in a real Cloud environment. The results show their feasibility and ability to reduce application costs, and to dynamically adjust the behavior at runtime without violating their correctness
    corecore