235 research outputs found

    On Single-Pushout Rewriting of Partial Algebras

    Get PDF
    We introduce Single-Pushout Rewriting for arbitrary partial algebras. Thus, we give up the usual restriction to graph structures, which are algebraic categories with unary operators only. By this generalisation, we obtain an integrated and straightforward treatment of graphical structures (objects) and attributes (data). We lose co-completeness of the underlying category. Therefore, a rule is no longer applicable at any match. We characterise the new application condition and make constructive use of it in some practical examples

    Piecewise linear approximations for the static-dynamic uncertainty strategy in stochastic lot-sizing

    Get PDF
    In this paper, we develop mixed integer linear programming models to compute near-optimal policy parameters for the non-stationary stochastic lot sizing problem under Bookbinder and Tan's static-dynamic uncertainty strategy. Our models build on piecewise linear upper and lower bounds of the first order loss function. We discuss different formulations of the stochastic lot sizing problem, in which the quality of service is captured by means of backorder penalty costs, non-stockout probability, or fill rate constraints. These models can be easily adapted to operate in settings in which unmet demand is backordered or lost. The proposed approach has a number of advantages with respect to existing methods in the literature: it enables seamless modelling of different variants of the above problem, which have been previously tackled via ad-hoc solution methods; and it produces an accurate estimation of the expected total cost, expressed in terms of upper and lower bounds. Our computational study demonstrates the effectiveness and flexibility of our models.Comment: 38 pages, working draf

    On the equivalence of strong formulations for capacitated multi-level lot sizing problems with setup times

    Get PDF
    Several mixed integer programming formulations have been proposed for modeling capacitated multi-level lot sizing problems with setup times. These formulations include the so-called facility location formulation, the shortest route formulation, and the inventory and lot sizing formulation with (l,S) inequalities. In this paper, we demonstrate the equivalence of these formulations when the integrality requirement is relaxed for any subset of binary setup decision variables. This equivalence has significant implications for decomposition-based methods since same optimal solution values are obtained no matter which formulation is used. In particular, we discuss the relax-and-fix method, a decomposition-based heuristic used for the efficient solution of hard lot sizing problems. Computational tests allow us to compare the effectiveness of different formulations using benchmark problems. The choice of formulation directly affects the required computational effort, and our results therefore provide guidelines on choosing an effective formulation during the development of heuristic-based solution procedures

    Enriching and validating geographic information on the web

    Get PDF
    The continuous growth of available data on the World Wide Web has led to an unprecedented amount of available information. However, the enormous variance in data quality and trustworthiness of information sources impairs the great potential of the large amount of vacant information. This observation especially applies to geographic information on the Web, i.e., information describing entities that are located on the Earth’s surface. With the advent of mobile devices, the impact of geographic Web information on our everyday life has substantially grown. The mobile devices have also enabled the creation of novel data sources such as OpenStreetMap (OSM), a collaborative crowd-sourced map providing open cartographic information. Today, we use geographic information in many applications, including routing, location recommendation, or geographic question answering. The processing of geographic Web information yields unique challenges. First, the descriptions of geographic entities on the Web are typically not validated. Since not all Web information sources are trustworthy, the correctness of some geographic Web entities is questionable. Second, geographic information sources on the Web are typically isolated from each other. The missing integration of information sources hinders the efficient use of geographic Web information for many applications. Third, the description of geographic entities is typically incomplete. Depending on the application, missing information is a decisive criterion for (not) using a particular data source. Due to the large scale of the Web, the manual correction of these problems is usually not feasible such that automated approaches are required. In this thesis, we tackle these challenges from three different angles. (i) Validation of geographic Web information: We validate geographic Web information by detecting vandalism in OpenStreetMap, for instance, the replacement of a street name with advertisement. To this end, we present the OVID model for automated vandalism detection in OpenStreetMap. (ii) Enrichment of geographic Web information through integration: We integrate OpenStreetMap with other geographic Web information sources, namely knowledge graphs, by identifying entries corresponding to the same world real-world entities in both data sources. We present the OSM2KG model for automated identity link discovery between OSM and knowledge graphs. (iii) Enrichment of missing information in geographic Web information: We consider semantic annotations of geographic entities on Web pages as an additional data source. We exploit existing annotations of categorical properties of Web entities as training data to enrich missing categorical properties in geographic Web information. For all of the proposed models, we conduct extensive evaluations on real-world datasets. Our experimental results confirm that the proposed solutions reliably outperform existing baselines. Furthermore, we demonstrate the utility of geographic Web Information in two application scenarios. (i) Corpus of geographic entity embeddings: We introduce the GeoVectors corpus, a linked open dataset of ready-to-use embeddings of geographic entities. With GeoVectors, we substantially lower the burden to use geographic data in machine learning applications. (ii) Application to event impact prediction: We employ several geographic Web information sources to predict the impact of public events on road traffic. To this end, we use cartographic, event, and event venue information from the Web.Durch die kontinuierliche Zunahme verfĂŒgbarer Daten im World Wide Web, besteht heute eine noch nie da gewesene Menge verfĂŒgbarer Informationen. Das große Potential dieser Daten wird jedoch durch hohe Schwankungen in der DatenqualitĂ€t und in der VertrauenswĂŒrdigkeit der Datenquellen geschmĂ€lert. Dies kann vor allem am Beispiel von geografischen Web-Informationen beobachtet werden. Geografische Web-Informationen sind Informationen ĂŒber EntitĂ€ten, die ĂŒber Koordinaten auf der ErdoberflĂ€che verfĂŒgen. Die Relevanz von geografischen Web-Informationen fĂŒr den Alltag ist durch die Verbreitung von internetfĂ€higen, mobilen EndgerĂ€ten, zum Beispiel Smartphones, extrem gestiegen. Weiterhin hat die VerfĂŒgbarkeit der mobilen EndgerĂ€te auch zur Erstellung neuartiger Datenquellen wie OpenStreetMap (OSM) gefĂŒhrt. OSM ist eine offene, kollaborative Webkarte, die von Freiwilligen dezentral erstellt wird. Mittlerweile ist die Nutzung geografischer Informationen die Grundlage fĂŒr eine Vielzahl von Anwendungen, wie zum Beispiel Navigation, Reiseempfehlungen oder geografische Frage-Antwort-Systeme. Bei der Verarbeitung geografischer Web-Informationen mĂŒssen einzigartige Herausforderungen berĂŒcksichtigt werden. Erstens werden die Beschreibungen geografischer Web-EntitĂ€ten typischerweise nicht validiert. Da nicht alle Informationsquellen im Web vertrauenswĂŒrdig sind, ist die Korrektheit der Darstellung mancher Web-EntitĂ€ten fragwĂŒrdig. Zweitens sind Informationsquellen im Web oft voneinander isoliert. Die fehlende Integration von Informationsquellen erschwert die effektive Nutzung von geografischen Web-Information in vielen AnwendungsfĂ€llen. Drittens sind die Beschreibungen von geografischen EntitĂ€ten typischerweise unvollstĂ€ndig. Je nach Anwendung kann das Fehlen von bestimmten Informationen ein entscheidendes Kriterium fĂŒr die Nutzung einer Datenquelle sein. Da die GrĂ¶ĂŸe des Webs eine manuelle Behebung dieser Probleme nicht zulĂ€sst, sind automatisierte Verfahren notwendig. In dieser Arbeit nĂ€hern wir uns diesen Herausforderungen von drei verschiedenen Richtungen. (i) Validierung von geografischen Web-Informationen: Wir validieren geografische Web-Informationen, indem wir Vandalismus in OpenStreetMap identifizieren, zum Beispiel das Ersetzen von Straßennamen mit Werbetexten. (ii) Anreicherung von geografischen Web-Information durch Integration: Wir integrieren OpenStreetMap mit anderen Informationsquellen im Web (Wissensgraphen), indem wir EintrĂ€ge in beiden Informationsquellen identifizieren, die den gleichen Echtwelt-EntitĂ€ten entsprechen. (iii) Anreicherung von fehlenden geografischen Informationen: Wir nutzen semantische Annotationen von geografischen EntitĂ€ten auf Webseiten als weitere Datenquelle. Wir nutzen existierende Annotationen kategorischer Attribute von Web-EntitĂ€ten als Trainingsdaten, um fehlende kategorische Attribute in geografischen Web-Informationen zu ergĂ€nzen. Wir fĂŒhren ausfĂŒhrliche Evaluationen fĂŒr alle beschriebenen Modelle durch. Die vorgestellten LösungsansĂ€tze erzielen verlĂ€sslich bessere Ergebnisse als existierende AnsĂ€tze. Weiterhin demonstrieren wir den Nutzen von geografischen Web-Informationen in zwei Anwendungsszenarien. (i) Korpus mit Embeddings von geografischen EntitĂ€ten: Wir stellen den GeoVectors-Korpus vor, einen verlinkten, offenen Datensatz mit direkt nutzbaren Embeddings von geografischen Web-EntitĂ€ten. Der GeoVectors-Korpus erleichtert die Nutzung von geografischen Daten in Anwendungen von maschinellen Lernen erheblich. (ii) Anwendung zur Prognose von Veranstaltungsauswirkungen: Wir nutzen Karten-, Veranstaltungs- und VeranstaltungsstĂ€tten-Daten aus dem Web, um die Auswirkungen von Veranstaltungen auf den Straßenverkehr zu prognostizieren

    Mining topological dependencies of recurrent congestion in road networks

    Get PDF
    The discovery of spatio-temporal dependencies within urban road networks that cause Recurrent Congestion (RC) patterns is crucial for numerous real-world applications, including urban planning and the scheduling of public transportation services. While most existing studies investigate temporal patterns of RC phenomena, the influence of the road network topology on RC is often over-looked. This article proposes the ST-DISCOVERY algorithm, a novel unsupervised spatio-temporal data mining algorithm that facilitates effective data-driven discovery of RC dependencies induced by the road network topology using real-world traffic data. We factor out regularly reoccurring traffic phenomena, such as rush hours, mainly induced by the daytime, by modelling and systematically exploiting temporal traffic load outliers. We present an algorithm that first constructs connected subgraphs of the road network based on the traffic speed outliers. Second, the algorithm identifies pairs of subgraphs that indicate spatio-temporal correlations in their traffic load behaviour to identify topological dependencies within the road network. Finally, we rank the identified subgraph pairs based on the dependency score determined by our algorithm. Our experimental results demonstrate that ST-DISCOVERY can effectively reveal topological dependencies in urban road networks

    A Detailed Report on the Overhead of Hardware APIs for Lightweight Cryptography

    Get PDF
    The Competition for Authenticated Encryption: Security, Applicability, and Robustness (CAESAR) was the first cryptographic competition that required designers to use a mandatory hardware API for their implementations. Recently, a similar hardware API for the NIST Lightweight Cryptography (LWC) project was proposed. Both APIs feature an accompanying development package to help designers implementing the API. In this paper, we have an in-depth look on these packages. We analyze the features of both packages, discuss their resource utilization, and demonstrate their impact on Ascon128, SpoC-64, and Gimli implementations on a modern Artix-7 FPGA. Finally, we provide some tweaks and enhancements to further optimize the development package for the LWC API
    • 

    corecore