467 research outputs found

    Parallel and Flow-Based High Quality Hypergraph Partitioning

    Get PDF
    Balanced hypergraph partitioning is a classic NP-hard optimization problem that is a fundamental tool in such diverse disciplines as VLSI circuit design, route planning, sharding distributed databases, optimizing communication volume in parallel computing, and accelerating the simulation of quantum circuits. Given a hypergraph and an integer kk, the task is to divide the vertices into kk disjoint blocks with bounded size, while minimizing an objective function on the hyperedges that span multiple blocks. In this dissertation we consider the most commonly used objective, the connectivity metric, where we aim to minimize the number of different blocks connected by each hyperedge. The most successful heuristic for balanced partitioning is the multilevel approach, which consists of three phases. In the coarsening phase, vertex clusters are contracted to obtain a sequence of structurally similar but successively smaller hypergraphs. Once sufficiently small, an initial partition is computed. Lastly, the contractions are successively undone in reverse order, and an iterative improvement algorithm is employed to refine the projected partition on each level. An important aspect in designing practical heuristics for optimization problems is the trade-off between solution quality and running time. The appropriate trade-off depends on the specific application, the size of the data sets, and the computational resources available to solve the problem. Existing algorithms are either slow, sequential and offer high solution quality, or are simple, fast, easy to parallelize, and offer low quality. While this trade-off cannot be avoided entirely, our goal is to close the gaps as much as possible. We achieve this by improving the state of the art in all non-trivial areas of the trade-off landscape with only a few techniques, but employed in two different ways. Furthermore, most research on parallelization has focused on distributed memory, which neglects the greater flexibility of shared-memory algorithms and the wide availability of commodity multi-core machines. In this thesis, we therefore design and revisit fundamental techniques for each phase of the multilevel approach, and develop highly efficient shared-memory parallel implementations thereof. We consider two iterative improvement algorithms, one based on the Fiduccia-Mattheyses (FM) heuristic, and one based on label propagation. For these, we propose a variety of techniques to improve the accuracy of gains when moving vertices in parallel, as well as low-level algorithmic improvements. For coarsening, we present a parallel variant of greedy agglomerative clustering with a novel method to resolve cluster join conflicts on-the-fly. Combined with a preprocessing phase for coarsening based on community detection, a portfolio of from-scratch partitioning algorithms, as well as recursive partitioning with work-stealing, we obtain our first parallel multilevel framework. It is the fastest partitioner known, and achieves medium-high quality, beating all parallel partitioners, and is close to the highest quality sequential partitioner. Our second contribution is a parallelization of an n-level approach, where only one vertex is contracted and uncontracted on each level. This extreme approach aims at high solution quality via very fine-grained, localized refinement, but seems inherently sequential. We devise an asynchronous n-level coarsening scheme based on a hierarchical decomposition of the contractions, as well as a batch-synchronous uncoarsening, and later fully asynchronous uncoarsening. In addition, we adapt our refinement algorithms, and also use the preprocessing and portfolio. This scheme is highly scalable, and achieves the same quality as the highest quality sequential partitioner (which is based on the same components), but is of course slower than our first framework due to fine-grained uncoarsening. The last ingredient for high quality is an iterative improvement algorithm based on maximum flows. In the sequential setting, we first improve an existing idea by solving incremental maximum flow problems, which leads to smaller cuts and is faster due to engineering efforts. Subsequently, we parallelize the maximum flow algorithm and schedule refinements in parallel. Beyond the strive for highest quality, we present a deterministically parallel partitioning framework. We develop deterministic versions of the preprocessing, coarsening, and label propagation refinement. Experimentally, we demonstrate that the penalties for determinism in terms of partition quality and running time are very small. All of our claims are validated through extensive experiments, comparing our algorithms with state-of-the-art solvers on large and diverse benchmark sets. To foster further research, we make our contributions available in our open-source framework Mt-KaHyPar. While it seems inevitable, that with ever increasing problem sizes, we must transition to distributed memory algorithms, the study of shared-memory techniques is not in vain. With the multilevel approach, even the inherently slow techniques have a role to play in fast systems, as they can be employed to boost quality on coarse levels at little expense. Similarly, techniques for shared-memory parallelism are important, both as soon as a coarse graph fits into memory, and as local building blocks in the distributed algorithm

    LIPIcs, Volume 261, ICALP 2023, Complete Volume

    Get PDF
    LIPIcs, Volume 261, ICALP 2023, Complete Volum

    Simulating The Impact of Emissions Control on Economic Productivity Using Particle Systems and Puff Dispersion Model

    Get PDF
    A simulation platform is developed for quantifying the change in productivity of an economy under passive and active emission control mechanisms. The program uses object-oriented programming to code a collection of objects resembling typical stakeholders in an economy. These objects include firms, markets, transportation hubs, and boids which are distributed over a 2D surface. Firms are connected using a modified Prim’s Minimum spanning tree algorithm, followed by implementation of an all-pair shortest path Floyd Warshall algorithm for navigation purposes. Firms use a non-linear production function for transformation of land, labor, and capital inputs to finished product. A GA-Vehicle Routing Problem with multiple pickups and drop-offs is implemented for efficient delivery of commodities across multiple nodes in the economy. Boids are autonomous agents which perform several functions in the economy including labor, consumption, renting, saving, and investing. Each boid is programmed with several microeconomic functions including intertemporal choice models, Hicksian and Marshallian demand function, and labor-leisure model. The simulation uses a Puff Dispersion model to simulate the advection and diffusion of emissions from point and mobile sources in the economy. A dose-response function is implemented to quantify depreciation of a Boid’s health upon contact with these emissions. The impact of emissions control on productivity and air quality is examined through a series of passive and active emission control scenarios. Passive control examines the impact of various shutdown times on economic productivity and rate of emissions exposure experienced by boids. The active control strategy examines the effects of acceptable levels of emissions exposure on economic productivity. The key findings on 7 different scenarios of passive and active emissions controls indicate that rate of productivity and consumption in an economy declines with increased scrutiny of emissions from point sources. In terms of exposure rates, the point sources may not be the primary source of average exposure rates, however they significantly impact the maximum exposure rate experienced by a boid. Tightening of emissions control also negatively impacts the transportation sector by reducing the asset utilization rate as well as reducing the total volume of goods transported across the economy

    Polyelectrolyte complexes embedding reduced graphite oxide

    Get PDF
    L'abstract è presente nell'allegato / the abstract is in the attachmen

    Towards full-scale autonomy for multi-vehicle systems planning and acting in extreme environments

    Get PDF
    Currently, robotic technology offers flexible platforms for addressing many challenging problems that arise in extreme environments. These problems’ nature enhances the use of heterogeneous multi-vehicle systems which can coordinate and collaborate to achieve a common set of goals. While such applications have previously been explored in limited contexts, long-term deployments in such settings often require an advanced level of autonomy to maintain operability. The success of planning and acting approaches for multi-robot systems are conditioned by including reasoning regarding temporal, resource and knowledge requirements, and world dynamics. Automated planning provides the tools to enable intelligent behaviours in robotic systems. However, whilst many planning approaches and plan execution techniques have been proposed, these solutions highlight an inability to consistently build and execute high-quality plans. Motivated by these challenges, this thesis presents developments advancing state-of-the-art temporal planning and acting to address multi-robot problems. We propose a set of advanced techniques, methods and tools to build a high-level temporal planning and execution system that can devise, execute and monitor plans suitable for long-term missions in extreme environments. We introduce a new task allocation strategy, called HRTA, that optimises the task distribution amongst the heterogeneous fleet, relaxes the planning problem and boosts the plan search. We implement the TraCE planner that enforces contingent planning considering propositional temporal and numeric constraints to deal with partial observability about the initial state. Our developments regarding robust plan execution and mission adaptability include the HLMA, which efficiently optimises the task allocation and refines the planning model considering the experience from robots’ previous mission executions. We introduce the SEA failure solver that, combined with online planning, overcomes unexpected situations during mission execution, deals with joint goals implementation, and enhances mission operability in long-term deployments. Finally, we demonstrate the efficiency of our approaches with a series of experiments using a new set of real-world planning domains.Engineering and Physical Sciences Research Council (EPSRC) grant EP/R026173/

    Applications

    Get PDF
    Volume 3 describes how resource-aware machine learning methods and techniques are used to successfully solve real-world problems. The book provides numerous specific application examples: in health and medicine for risk modelling, diagnosis, and treatment selection for diseases in electronics, steel production and milling for quality control during manufacturing processes in traffic, logistics for smart cities and for mobile communications
    • …
    corecore