12,612 research outputs found

    Lying Your Way to Better Traffic Engineering

    Full text link
    To optimize the flow of traffic in IP networks, operators do traffic engineering (TE), i.e., tune routing-protocol parameters in response to traffic demands. TE in IP networks typically involves configuring static link weights and splitting traffic between the resulting shortest-paths via the Equal-Cost-MultiPath (ECMP) mechanism. Unfortunately, ECMP is a notoriously cumbersome and indirect means for optimizing traffic flow, often leading to poor network performance. Also, obtaining accurate knowledge of traffic demands as the input to TE is elusive, and traffic conditions can be highly variable, further complicating TE. We leverage recently proposed schemes for increasing ECMP's expressiveness via carefully disseminated bogus information ("lies") to design COYOTE, a readily deployable TE scheme for robust and efficient network utilization. COYOTE leverages new algorithmic ideas to configure (static) traffic splitting ratios that are optimized with respect to all (even adversarially chosen) traffic scenarios within the operator's "uncertainty bounds". Our experimental analyses show that COYOTE significantly outperforms today's prevalent TE schemes in a manner that is robust to traffic uncertainty and variation. We discuss experiments with a prototype implementation of COYOTE

    New directions for Artificial Intelligence (AI) methods in optimum design

    Get PDF
    Developments and applications of artificial intelligence (AI) methods in the design of structural systems is reviewed. Principal shortcomings in the current approach are emphasized, and the need for some degree of formalism in the development environment for such design tools is underscored. Emphasis is placed on efforts to integrate algorithmic computations in expert systems

    SUBIC: A Supervised Bi-Clustering Approach for Precision Medicine

    Full text link
    Traditional medicine typically applies one-size-fits-all treatment for the entire patient population whereas precision medicine develops tailored treatment schemes for different patient subgroups. The fact that some factors may be more significant for a specific patient subgroup motivates clinicians and medical researchers to develop new approaches to subgroup detection and analysis, which is an effective strategy to personalize treatment. In this study, we propose a novel patient subgroup detection method, called Supervised Biclustring (SUBIC) using convex optimization and apply our approach to detect patient subgroups and prioritize risk factors for hypertension (HTN) in a vulnerable demographic subgroup (African-American). Our approach not only finds patient subgroups with guidance of a clinically relevant target variable but also identifies and prioritizes risk factors by pursuing sparsity of the input variables and encouraging similarity among the input variables and between the input and target variable

    Automatic map-based FTTx access network design

    Get PDF
    Several mature and standardized optical access network technologies are available for network operators providing broadband services, being now in deployment phase; therefore cost estimation, business analysis, efficient deployment strategies, network and topology design issues for FTTx access networks play an increasingly important role regarding profitability and market success. In a competitive environment, techno-economic evaluation supports the optimal choice among available technologies. Even the tradeoff between future proof technical superiority and short term investment minimization requires a farseeing decision. In our point of view, cost estimation and techno-economic evaluation is strongly related to strategic network design: among others the uneven population density, irregular street system or infrastructure have significant impact on the network topology, thus the deployment costs as well. In order to deal with these aspects, a high-level, strategic network design is necessary that adapts to geospatial characteristics of the services area, providing accurate and detailed network information for the techno-economic evaluation [1]. We have developed a topology designer methodology that supprts the above requirements, providing (near) optimal topology of the fully or partially optical access network, based on the geospatial information about the service area: digital maps, existing infrastructure and subscriber database. Automatic topology design for large-scale service areas, with 10.000s of subsribers is a highly complex mathematical problem. The tough algorithms for a near optimal, yet efficient solution. The developed algorithms were evaluated regarding their speed and accuracy. Based on topology design results, a detailed and flexible techno-economic comparison is carried out, since the framework handles various broadband access network technologies, as presented in a case study. --Topology design,Strategic Design,Network planning,GIS,Map,Techno-economic,Cost estimation

    Dynamics of heuristic optimization algorithms on random graphs

    Full text link
    In this paper, the dynamics of heuristic algorithms for constructing small vertex covers (or independent sets) of finite-connectivity random graphs is analysed. In every algorithmic step, a vertex is chosen with respect to its vertex degree. This vertex, and some environment of it, is covered and removed from the graph. This graph reduction process can be described as a Markovian dynamics in the space of random graphs of arbitrary degree distribution. We discuss some solvable cases, including algorithms already analysed using different techniques, and develop approximation schemes for more complicated cases. The approximations are corroborated by numerical simulations.Comment: 19 pages, 3 figures, version to app. in EPJ

    Algorithmic and Statistical Perspectives on Large-Scale Data Analysis

    Full text link
    In recent years, ideas from statistics and scientific computing have begun to interact in increasingly sophisticated and fruitful ways with ideas from computer science and the theory of algorithms to aid in the development of improved worst-case algorithms that are useful for large-scale scientific and Internet data analysis problems. In this chapter, I will describe two recent examples---one having to do with selecting good columns or features from a (DNA Single Nucleotide Polymorphism) data matrix, and the other having to do with selecting good clusters or communities from a data graph (representing a social or information network)---that drew on ideas from both areas and that may serve as a model for exploiting complementary algorithmic and statistical perspectives in order to solve applied large-scale data analysis problems.Comment: 33 pages. To appear in Uwe Naumann and Olaf Schenk, editors, "Combinatorial Scientific Computing," Chapman and Hall/CRC Press, 201

    Some recent results in the analysis of greedy algorithms for assignment problems

    Get PDF
    We survey some recent developments in the analysis of greedy algorithms for assignment and transportation problems. We focus on the linear programming model for matroids and linear assignment problems with Monge property, on general linear programs, probabilistic analysis for linear assignment and makespan minimization, and on-line algorithms for linear and non-linear assignment problems
    • …
    corecore