61 research outputs found

    New Results on the Probabilistic Analysis of Online Bin Packing and its Variants

    Get PDF
    The classical bin packing problem can be stated as follows: We are given a multiset of items {a1, ..., an} with sizes in [0,1], and want to pack them into a minimum number of bins, each of which with capacity one. There are several applications of this problem, for example in the field of logistics: We can interpret the i-th item as time a package deliverer spends for the i-th tour. Package deliverers have a daily restricted working time, and we want to assign the tours such that the number of package deliverers needed is minimized. Another setup is to think of the items as boxes with a standardized basis, but variable height. Then, the goal is to pack these boxes into a container, which is standardized in all three dimensions. Moreover, applications of variants of the classical bin packing problem arise in cloud computing, when we have to store virtual machines on servers. Besides its practical relevance, the bin packing problem is one of the fundamental problems in theoretical computer science: It was proven many years ago that under standard complexity assumptions it is not possible to compute the value of an optimal packing of the items efficiently - classical bin packing is NP-complete. Computing the value efficiently means that the runtime of the algorithm is bounded polynomially in the number of items we have to pack. Besides the offline version, where we know all items at the beginning, also the online version is of interest: Here, the items are revealed one-by-one and have to be packed into a bin immediately and irrevocably without knowing which and how many items will still arrive in the future. Also this version is of practical relevance. In many situations we do not know the whole input at the beginning: For example we are unaware of the requirements of future virtual machines, which have to be stored, or suddenly some more packages have to be delivered, and some deliverers already started their tour. We can think of the classical theoretical analysis of an online algorithm A as follows: An adversary studies the behavior of the algorithm and afterwards constructs a sequence of items I. Then, the performance is measured by the number of used bins by A performing on I, divided by the value of an optimal packing of the items in I. The adversary tries to choose a worst-case sequence so this way to measure the performance is very pessimistic. Moreover, the chosen sequences I often turn out to be artificial: For example, in many cases the sizes of the items increase monotonically over time. Instances in practice are often subject to random influence and therefore it is likely that they are good-natured. In this thesis we analyze the performance of online algorithms with respect to two stochastic models. The first model is the following: The adversary chooses a set of items SI and a distribution F on SI. Then, the items are drawn independently and identically distributed according to F. In the second model the adversary chooses a finite set of items SI and then these items arrive in random order, that is random with respect to the uniform distribution on the set of all possible permutations of the items. It is possible to show that the adversary in the second stochastic model is at least as powerful as in the first one. We can classify the results in this thesis in three parts: In the first part we consider the complexity of classical bin packing and its variants cardinality-constrained and class-constrained bin packing in both stochastic models. That is, we determine if it is possible to construct algorithms that are in expectation nearly optimal for large instances that are constructed according to the stochastic models or if there exist non-trivial lower bounds. Among other things we show that the complexity of class-constrained bin packing differs in the two models under consideration. In the second part we deal with bounded-space bin packing and the dual maximization variant bin covering. We show that it is possible to overcome classical worst-case bounds in both models. In other words, we see that bounded-space algorithms benefit from randomized instances compared to the worst case. Finally, we consider selected heuristics for class-constrained bin packing and the corresponding maximization variant class-constrained bin covering. Here, we note that the different complexity of class-constrained bin packing with respect to the studied stochastic models observed in the first part is not only a theoretical phenomenon, but also takes place for many common algorithmic approaches. Interestingly, when we apply the same algorithmic ideas to class-constrained bin covering, we benefit from both types of randomization similarly. </ul

    On Discrete Hyperbox Packing

    Get PDF
    Bin packing is a very important and popular research area in the computer science field. Past work showed many good and real-world packing algorithms. How- ever, due to the complexity of the problem in multiple-dimensional bin packing, also called hyperbox packing, we need more practical packing algorithms for its real-world applications. In this dissertation, we extend 1D packing algorithms to hyperbox packing prob- lems via a general framework that takes two inputs of a 1D packing algorithm and an instance of hyperbox packing problem and outputs a hyperbox packing algorithm. The extension framework significantly enriches the family of hyperbox-packing algorithms, generates many framework-based algorithms, and simultaneously calls for the analysis for those algorithms. We also analyze the performance of a couple of framework-based algorithms from two perspectives of worst-case performance and average-case performance. In worst- case analysis, we use the worst-case performance ratio as our metric and analyze the relationship of the ratio of framework-based algorithms and that of the corresponding 1D algorithms. We also compare their worst-case performance against two baselines: strip optimal algorithms and optimal algorithms. In average-case analysis, we use expected waste as a metric, analyze the waste of optimal hyperbox packing algorithms, and estimate the asymptotic forms of the waste for framework-based algorithms

    On the Application of PSpice for Localised Cloud Security

    Get PDF
    The work reported in this thesis commenced with a review of methods for creating random binary sequences for encoding data locally by the client before storing in the Cloud. The first method reviewed investigated evolutionary computing software which generated noise-producing functions from natural noise, a highly-speculative novel idea since noise is stochastic. Nevertheless, a function was created which generated noise to seed chaos oscillators which produced random binary sequences and this research led to a circuit-based one-time pad key chaos encoder for encrypting data. Circuit-based delay chaos oscillators, initialised with sampled electronic noise, were simulated in a linear circuit simulator called PSpice. Many simulation problems were encountered because of the nonlinear nature of chaos but were solved by creating new simulation parts, tools and simulation paradigms. Simulation data from a range of chaos sources was exported and analysed using Lyapunov analysis and identified two sources which produced one-time pad sequences with maximum entropy. This led to an encoding system which generated unlimited, infinitely-long period, unique random one-time pad encryption keys for plaintext data length matching. The keys were studied for maximum entropy and passed a suite of stringent internationally-accepted statistical tests for randomness. A prototype containing two delay chaos sources initialised by electronic noise was produced on a double-sided printed circuit board and produced more than 200 Mbits of OTPs. According to Vladimir Kotelnikov in 1941 and Claude Shannon in 1945, one-time pad sequences are theoretically-perfect and unbreakable, provided specific rules are adhered to. Two other techniques for generating random binary sequences were researched; a new circuit element, memristance was incorporated in a Chua chaos oscillator, and a fractional-order Lorenz chaos system with order less than three. Quantum computing will present many problems to cryptographic system security when existing systems are upgraded in the near future. The only existing encoding system that will resist cryptanalysis by this system is the unconditionally-secure one-time pad encryption

    Complexity, Emergent Systems and Complex Biological Systems:\ud Complex Systems Theory and Biodynamics. [Edited book by I.C. Baianu, with listed contributors (2011)]

    Get PDF
    An overview is presented of System dynamics, the study of the behaviour of complex systems, Dynamical system in mathematics Dynamic programming in computer science and control theory, Complex systems biology, Neurodynamics and Psychodynamics.\u

    Online Optimization with Lookahead

    Get PDF
    The main contributions of this thesis consist of the development of a systematic groundwork for comprehensive performance evaluation of algorithms in online optimization with lookahead and the subsequent validation of the presented approaches in theoretical analysis and computational experiments

    Preventing premature convergence and proving the optimality in evolutionary algorithms

    Get PDF
    http://ea2013.inria.fr//proceedings.pdfInternational audienceEvolutionary Algorithms (EA) usually carry out an efficient exploration of the search-space, but get often trapped in local minima and do not prove the optimality of the solution. Interval-based techniques, on the other hand, yield a numerical proof of optimality of the solution. However, they may fail to converge within a reasonable time due to their inability to quickly compute a good approximation of the global minimum and their exponential complexity. The contribution of this paper is a hybrid algorithm called Charibde in which a particular EA, Differential Evolution, cooperates with a Branch and Bound algorithm endowed with interval propagation techniques. It prevents premature convergence toward local optima and outperforms both deterministic and stochastic existing approaches. We demonstrate its efficiency on a benchmark of highly multimodal problems, for which we provide previously unknown global minima and certification of optimality

    Using MapReduce Streaming for Distributed Life Simulation on the Cloud

    Get PDF
    Distributed software simulations are indispensable in the study of large-scale life models but often require the use of technically complex lower-level distributed computing frameworks, such as MPI. We propose to overcome the complexity challenge by applying the emerging MapReduce (MR) model to distributed life simulations and by running such simulations on the cloud. Technically, we design optimized MR streaming algorithms for discrete and continuous versions of Conway’s life according to a general MR streaming pattern. We chose life because it is simple enough as a testbed for MR’s applicability to a-life simulations and general enough to make our results applicable to various lattice-based a-life models. We implement and empirically evaluate our algorithms’ performance on Amazon’s Elastic MR cloud. Our experiments demonstrate that a single MR optimization technique called strip partitioning can reduce the execution time of continuous life simulations by 64%. To the best of our knowledge, we are the first to propose and evaluate MR streaming algorithms for lattice-based simulations. Our algorithms can serve as prototypes in the development of novel MR simulation algorithms for large-scale lattice-based a-life models.https://digitalcommons.chapman.edu/scs_books/1014/thumbnail.jp
    • …
    corecore