171 research outputs found

    Candidate selection and instance ordering for realtime algorithm configuration

    Get PDF
    Many modern combinatorial solvers have a variety of parameters through which a user can customise their behaviour. Algorithm configuration is the process of selecting good values for these parameters in order to improve performance. Time and again algorithm configuration has been shown to significantly improve the performance of many algorithms for solving challenging computational problems. Automated systems for tuning parameters regularly out-perform human experts, sometimes but orders of magnitude. Online algorithm configurators, such as ReACTR, are able to tune a solver online without incurring costly offline training. As such ReACTR’s main focus is on runtime minimisation while solving combinatorial problems. To do this ReACTR adopts a one-pass methodology where each instance in a stream of instances to be solved is considered only as it arrives. As such ReACTR’s performance is sensitive to the order in which instances arrive. It is still not understood which instance orderings positively or negatively effect the performance of ReACTR. This paper investigates the effect of instance ordering and grouping by empirically evaluating different instance orderings based on difficulty and feature values. Though the end use is generally unable to control the order in which instances arrive it is important to understand which orderings impact Re- ACTR’s performance and to what extent. This study also has practical benefit as such orderings can occur organically. For example as business grows the problems it may encounter, such as routing or scheduling, often grow in size and difficulty. ReACTR’s performance also depends strongly configuration selection procedure used. This component controls which configurations are selected to run in parallel from the internal configuration pool. This paper evaluates various ranking mechanisms and different ways of combining them to better understand how the candidate selection procedure affects realtime algorithm configuration. We show that certain selection procedures are superior to others and that the order which instances arrive in determines which selection procedure performs best. We find that both instance order and grouping can significantly affect the overall solving time of the online automatic algorithm configurator ReACTR. One of the more surprising discoveries is that having groupings of similar instances can actually negatively impact on the overall performance of the configurator. In particular we show that orderings based on nearly any instance feature values can lead to significant reductions in total runtime over random instance orderings. In addition, certain candidate selection procedures are more suited to certain orderings than others and selecting the correct one can show a marked improvement in solving times

    Capping methods for the automatic configuration of optimization algorithms

    Get PDF
    Automatic configuration techniques are widely and successfully used to find good parameter settings for optimization algorithms. Configuration is costly, because it is necessary to evaluate many configurations on different instances. For decision problems, when the objective is to minimize the running time of the algorithm, many configurators implement capping methods to discard poor configurations early. Such methods are not directly applicable to optimization problems, when the objective is to optimize the cost of the best solution found, given a predefined running time limit. We propose new capping methods for the automatic configuration of optimization algorithms. They use the previous executions to determine a performance envelope, which is used to evaluate new executions and cap those that do not satisfy the envelope conditions. We integrate the capping methods into the irace configurator and evaluate them on different optimization scenarios. Our results show that the proposed methods can save from about 5% to 78% of the configuration effort, while finding configurations of the same quality. Based on the computational analysis, we identify two conservative and two aggressive methods, that save an average of about 20% and 45% of the configuration effort, respectively. We also provide evidence that capping can help to better use the available budget in scenarios with a configuration time limit.This research has been supported by Coordenação de Aperfeiçoa-mento de Pessoal de NĂ­vel Superior – Brasil (CAPES) – Finance Code001. M. de Souza acknowledges the support of the Santa Catarina State University, Brasil. M. Ritt acknowledges the support of CNPq, Brasil (grant 437859/2018-5) and Google Research Latin America (grant25111). M. LĂłpez-Ibåñez is a ‘‘Beatriz Galindo’’ Senior Distinguished Researcher (BEAGAL 18/00053) funded by the Spanish Ministry of Science and Innovation (MICINN). This research is partially funded by TAILOR ICT-48 Network (No 952215) funded by EU Horizon 2020 research and innovation programme. Funding for open access charge: Universidad de MĂĄlaga / CBU

    Real-time algorithm configuration

    Get PDF
    This dissertation presents a number of contributions to the field of algorithm configur- ation. In particular, we present an extension to the algorithm configuration problem, real-time algorithm configuration, where configuration occurs online on a stream of instances, without the need for prior training, and problem solutions are returned in the shortest time possible. We propose a framework for solving the real-time algorithm configuration problem, ReACT. With ReACT we demonstrate that by using the parallel computing architectures, commonplace in many systems today, and a robust aggregate ranking system, configuration can occur without any impact on performance from the perspective of the user. This is achieved by means of a racing procedure. We show two concrete instantiations of the framework, and show them to be on a par with or even exceed the state-of-the-art in offline algorithm configuration using empirical evaluations on a range of combinatorial problems from the literature. We discuss, assess, and provide justification for each of the components used in our framework instantiations. Specifically, we show that the TrueSkill ranking system commonly used to rank players’ skill in multiplayer games can be used to accurately es- timate the quality of an algorithm’s configuration using only censored results from races between algorithm configurations. We confirm that the order that problem instances arrive in influences the configuration performance and that the optimal selection of configurations to participate in races is dependent on the distribution of the incoming in- stance stream. We outline how to maintain a pool of quality configurations by removing underperforming configurations, and techniques to generate replacement configurations with minimal computational overhead. Finally, we show that the configuration space can be reduced using feature selection techniques from the machine learning literature, and that doing so can provide a boost in configuration performance

    Systems for AutoML Research

    Get PDF

    Working Notes from the 1992 AAAI Workshop on Automating Software Design. Theme: Domain Specific Software Design

    Get PDF
    The goal of this workshop is to identify different architectural approaches to building domain-specific software design systems and to explore issues unique to domain-specific (vs. general-purpose) software design. Some general issues that cut across the particular software design domain include: (1) knowledge representation, acquisition, and maintenance; (2) specialized software design techniques; and (3) user interaction and user interface

    RĂ©agir et s’adapter Ă  son environnement: Concevoir des mĂ©thodes autonomes pour l’optimisation combinatoire Ă  plusieurs objectifs

    Get PDF
    Large-scale optimisation problems are usually hard to solve optimally. Approximation algorithms such as metaheuristics, able to quickly find sub-optimal solutions, are often preferred. This thesis focuses on multi-objective local search (MOLS) algorithms, metaheuristics able to deal with the simultaneous optimisation of multiple criteria. As many algorithms, metaheuristics expose many parameters that significantly impact their performance. These parameters can be either predicted and set before the execution of the algorithm, or dynamically modified during the execution itself.While in the last decade many advances have been made on the automatic design of algorithms, the great majority of them only deal with single-objective algorithms and the optimisation of a single performance indicator such as the algorithm running time or the final solution quality. In this thesis, we investigate the relations between automatic algorithm design and multi-objective optimisation, with an application on MOLS algorithms.We first review possible MOLS strategies ans parameters and present a general, highly configurable, MOLS framework. We also propose MO-ParamILS, an automatic configurator specifically designed to deal with multiple performance indicators. Then, we conduct several studies on the automatic offline design of MOLS algorithms on multiple combinatorial bi-objective problems. Finally, we discuss two online extensions of classical algorithm configuration: first the integration of parameter control mechanisms, to benefit from having multiple configuration predictions; then the use of configuration schedules, to sequentially use multiple configurations.Les problĂšmes d’optimisation Ă  grande Ă©chelle sont gĂ©nĂ©ralement difficiles Ă  rĂ©soudre de façon optimale. Des algorithmes d’approximation tels que les mĂ©taheuristiques, capables de trouver rapidement des solutions sous-optimales, sont souvent prĂ©fĂ©rĂ©s. Cette thĂšse porte sur les algorithmes de recherche locale multi-objectif (MOLS), des mĂ©taheuristiques capables de traiter l’optimisation simultanĂ©e de plusieurs critĂšres. Comme de nombreux algorithmes, les MOLS exposent de nombreux paramĂštres qui ont un impact important sur leurs performances. Ces paramĂštres peuvent ĂȘtre soit prĂ©dits et dĂ©finis avant l’exĂ©cution de l’algorithme, soit ensuite modifiĂ©s dynamiquement.Alors que de nombreux progrĂšs ont rĂ©cemment Ă©tĂ© rĂ©alisĂ©s pour la conception automatique d’algorithmes, la grande majoritĂ© d’entre eux ne traitent que d’algorithmes mono-objectif et l’optimisation d’un unique indicateur de performance. Dans cette thĂšse, nous Ă©tudions les relations entre la conception automatique d’algorithmes et l’optimisation multi-objective.Nous passons d’abord en revue les stratĂ©gies MOLS possibles et prĂ©sentons un framework MOLS gĂ©nĂ©ral et hautement configurable. Nous proposons Ă©galement MO-ParamILS, un configurateur automatique spĂ©cialement conçu pour gĂ©rer plusieurs indicateurs de performance. Nous menons ensuite plusieurs Ă©tudes sur la conception automatique de MOLS sur de multiples problĂšmes combinatoires bi-objectifs. Enfin, nous discutons deux extensions de la configuration d’algorithme classique : d’abord l’intĂ©gration des mĂ©canismes de contrĂŽle de paramĂštres, pour bĂ©nĂ©ficier de multiples prĂ©dictions de configuration; puis l’utilisation sĂ©quentielle de plusieurs configurations

    Providing machine level data for cloud based analytics

    Get PDF
    The target of this thesis was to investigate the problems, possibilities and background of transporting data collected in a manufacturing process to a cloud environment for analytics. The underlying reason for this was to spot themes and topics that need to be addressed when planning a real-life project aiming to improve manufacturing with cloud manufacturing. To achieve this, first the underlying theories were studied to understand why we actually need cloud analytics and what kind of paradigms could help in achieving it. The reasoning was mostly based on themes revolving around Germany’s Industry 4.0 initiative. After that, the state of the art was researched to understand what has been already done and how some fundamental problems have been solved. From this base, requirements for a solution model were made. After requirements for a good system were formulated, some of the most promising technologies were researched and after that evaluated according to the requirements. This thesis was able to highlight many core problems that modern cloud analytic systems for manufacturing will face. The most central technological findings revolve around a vital need for industry wide standardization, which is a theme that requires both customers and vendors: customers need to demand vendors to embrace common open standards and vendors need to respond to this need. Another core problem discussed in this thesis was the usage of different cloud services. Many manufacturing giants offer their own platforms for providing cloud analytics, but also bring some fundamental problems to play. On the other hand open cloud platform providers like Microsoft Azure and Amazon Web Services offer a very large offering of components, services and applications with competitive pricing. Altogether this thesis produced some topics that require a thorough analysis when planning a system like this and some insight into the core problems. In addition, some recommendations were produced on how to tackle problems in ways that transition well in to the future as the field is evolving rapidly and future compatibility is important.TĂ€mĂ€n diplomityön tarkoituksena oli tutkia datan kerĂ€ystĂ€ tuotantoprosessista pilviympĂ€ristöön analytiikkaa varten, sekĂ€ siihen liittyviĂ€ ilmiöitĂ€, ongelmia sekĂ€ mahdollisuuksia. Motivaatioina tĂ€hĂ€n kaikkeen oli havaita mahdollisia aiheita ja aihepiirejĂ€, joita tulisi tutkia ja arvioida reaalimaailman projektia suunnitellessa. Tavoitteisiin pÀÀsemiseksi aluksi tutkittiin teorioita, joiden avulla haettiin ymmĂ€rrystĂ€ miksi oikeastaan pilvianalytiikkaa tarvitaan ja minkĂ€laiset ratkaisumallit voisivat auttaa sen toteuttamisessa. KĂ€sitellyt teemat liittyivĂ€t hyvin vahvasti Saksan Industry 4.0 hankkeeseen ja siihen liittyvÀÀn tutkimustyöhön. TĂ€mĂ€n jĂ€lkeen tarkasteltiin nykytilaa ja minkĂ€laisilla ratkaisuilla ongelmaa yleensĂ€ on ratkaistu. TĂ€mĂ€n pohjalta muotoiltiin vaatimuksia ratkaisumallille. VaatimusmÀÀrittelyiden pohjalta alettiin tutkia lupaavimpia teknologioita ja tekniikoita vaatimusten pohjalta. Työ onnistui nostamaan esille useita keskeisiĂ€ ongelmia aiheuttavia teemoja, mitĂ€ modernit pilvianalytiikkaratkaisut joutuvat kohtaamaan. KeskeisimmĂ€t tekniikkaan liittyvĂ€t löydökset ovat vahvasti kytköksissĂ€ toimialojen laajuisen standardoinnin tarpeeseen. TĂ€mĂ€ on teema, joka vaatii toimenpiteitĂ€ niin asiakkailta kuin toimittajilta: asiakkaiden tulee vaatia toimittajia perustamaan ratkaisunsa avoimille ja yhteisille standardeille ja toimittajien tulee vastata tĂ€hĂ€n tarpeeseen. Toinen keskeinen kĂ€sitelty ongelma oli erilaisten pilvipalveluiden vertailu ja niistĂ€ oikean valinta. Useat valmistavan teollisuuden toimittajajĂ€tit tarjoavat omia alustoja pilvianalytiikalle, jotka tarjoavat valmiita ratkaisuja ainakin joihinkin ongelmiin. Toisaalta avoimet pilvitoimittajat kuten Microsoft (Azure) ja Amazon (AWS) tarjoavat erittĂ€in laajalla skaalalla erilaisia komponentteja, palveluita ja sovelluksia kilpailukykyisellĂ€ hinnoittelulla. Yhteenvetona tĂ€mĂ€ diplomityö nosti ja tuotti pohdintaa useasta eri aihealueesta, mitkĂ€ vaativat huolellista analyysiĂ€ reaalimaailman ratkaisua suunnitellessa. LisĂ€ksi pohdinnan perusteella pystyttiin muotoilemaan joitakin ratkaisuehdotuksia ongelmien vĂ€lttĂ€miseksi tavalla, jotka tulevaisuudessa tukevat kĂ€sitellyn alan nopeasti kehittyvÀÀ luonnetta ja korkeita vaatimuksia yhteensopivuudelle

    Scalable and Reliable Middlebox Deployment

    Get PDF
    Middleboxes are pervasive in modern computer networks providing functionalities beyond mere packet forwarding. Load balancers, intrusion detection systems, and network address translators are typical examples of middleboxes. Despite their benefits, middleboxes come with several challenges with respect to their scalability and reliability. The goal of this thesis is to devise middlebox deployment solutions that are cost effective, scalable, and fault tolerant. The thesis includes three main contributions: First, distributed service function chaining with multiple instances of a middlebox deployed on different physical servers to optimize resource usage; Second, Constellation, a geo-distributed middlebox framework enabling a middlebox application to operate with high performance across wide area networks; Third, a fault tolerant service function chaining system
    • 

    corecore