69 research outputs found
Online Bin Stretching with Three Bins
Online Bin Stretching is a semi-online variant of bin packing in which the
algorithm has to use the same number of bins as an optimal packing, but is
allowed to slightly overpack the bins. The goal is to minimize the amount of
overpacking, i.e., the maximum size packed into any bin.
We give an algorithm for Online Bin Stretching with a stretching factor of
for three bins. Additionally, we present a lower bound of for Online Bin Stretching on three bins and a lower bound of
for four and five bins that were discovered using a computer search.Comment: Preprint of a journal version. See version 2 for the conference
paper. Conference paper split into two journal submissions; see
arXiv:1601.0811
On packet scheduling with adversarial jamming and speedup
In Packet Scheduling with Adversarial Jamming, packets of arbitrary sizes arrive over time to be transmitted over a channel in which instantaneous jamming errors occur at times chosen by the adversary and not known to the algorithm. The transmission taking place at the time of jamming is corrupt, and the algorithm learns this fact immediately. An online algorithm maximizes the total size of packets it successfully transmits and the goal is to develop an algorithm with the lowest possible asymptotic competitive ratio, where the additive constant may depend on packet sizes. Our main contribution is a universal algorithm that works for any speedup and packet sizes and, unlike previous algorithms for the problem, it does not need to know these parameters in advance. We show that this algorithm guarantees 1-competitiveness with speedup 4, making it the first known algorithm to maintain 1-competitiveness with a moderate speedup in the general setting of arbitrary packet sizes. We also prove a lower bound of ϕ+1≈2.618 on the speedup of any 1-competitive deterministic algorithm, showing that our algorithm is close to the optimum. Additionally, we formulate a general framework for analyzing our algorithm locally and use it to show upper bounds on its competitive ratio for speedups in [1, 4) and for several special cases, recovering some previously known results, each of which had a dedicated proof. In particular, our algorithm is 3-competitive without speedup, matching both the (worst-case) performance of the algorithm by Jurdzinski et al. (Proceedings of the 12th workshop on approximation and online algorithms (WAOA), LNCS 8952, pp 193–206, 2015. http://doi.org/10.1007/978-3-319-18263-6_17) and the lower bound by Anta et al. (J Sched 19(2):135–152, 2016. http://doi.org/10.1007/s10951-015-0451-z)
A Time-Series Compression Technique and its Application to the Smart Grid
Time-series data is increasingly collected in many domains. One example is the smart electricity infrastructure, which generates huge volumes of such data from sources such as smart electricity meters. Although today this data is used for visualization and billing in mostly 15-min resolution, its original temporal resolution frequently is more fine-grained, e.g., seconds. This is useful for various analytical applications such as short-term forecasting, disaggregation and visualization. However, transmitting and storing huge amounts of such fine-grained data is prohibitively expensive in terms of storage space in many cases. In this article, we present a compression technique based on piecewise regression and two methods which describe the performance of the compression. Although our technique is a general approach for time-series compression, smart grids serve as our running example and as our evaluation scenario. Depending on the data and the use-case scenario, the technique compresses data by ratios of up to factor 5,000 while maintaining its usefulness for analytics. The proposed technique has outperformed related work and has been applied to three real-world energy datasets in different scenarios. Finally, we show that the proposed compression technique can be implemented in a state-of-the-art database management system
Comparison of Pandemics and Epidemics of the 20th and 21st Centuries
Od 12. března 2020 až po současnost jsou v České republice
zavedeny v určité míře protiepidemická opatření z důvodu omezení
šíření SARS-CoV-2. Šíření tvz. „nového“ druhu lidského koronaviru
je velmi podrobně sledováno a vyhodnocováno na celé planetě.
Jsou zaváděna opatření proti jeho šíření a dopadům na společnost
dle aktuálního stavu v daném státu. Kroky ke snížení šíření nákazy
byly zaváděny i při minulých epidemiích a pandemiích.
Článek se zabývá porovnáním proběhlých a probíhajících
epidemií a pandemií 20. a 21. století.
Cílem je komparace epidemií a pandemií 20. a 21. století
s probíhající pandemií COVID-19. V rámci vlastního šetření jsou
vybrány významné pandemie podle počtu mrtvých, ke kterým byly
vytvořeny přehledové grafy. Z šetření vyplývá, že viry jsou původci
všech porovnávaných pandemií a dominuje jim přenos kapénkami.
Dále vyplývá vysoká podobnost mezi pandemií španělské chřipky
a pandemií COVID-19.From 12 March 2020 to the present day, certain antiepidemic
measures have been introduced in the Czech Republic to limit
the spread of SARS-CoV-2. The spread of the so-called. The spread
of the "new" type of human coronavirus is being monitored and
evaluated very closely across the planet. Measures against its
spread and its impact on society are being implemented according
to the current situation in the country concerned. Steps to reduce
the spread of the disease have also been implemented in past
epidemics and pandemics.
The article deals with a comparison of past and ongoing
epidemics and pandemics of the 20th and 21st centuries.
The aim is to compare the epidemics and pandemics of
the 20th and 21st centuries with the ongoing pandemic COVID-19.
In the actual investigation, major pandemics are selected according
to the number of casualties for which overview graphs have been
produced. The investigation shows that viruses are the causative
agents of the pandemics compared and are dominated by droplet
transmission. It also shows a high similarity between the Spanish
influenza pandemic and the COVID-19 pandemic
Online Algorithms for Multi-Level Aggregation
In the Multi-Level Aggregation Problem (MLAP), requests arrive at the nodes
of an edge-weighted tree T, and have to be served eventually. A service is
defined as a subtree X of T that contains its root. This subtree X serves all
requests that are pending in the nodes of X, and the cost of this service is
equal to the total weight of X. Each request also incurs waiting cost between
its arrival and service times. The objective is to minimize the total waiting
cost of all requests plus the total cost of all service subtrees. MLAP is a
generalization of some well-studied optimization problems; for example, for
trees of depth 1, MLAP is equivalent to the TCP Acknowledgment Problem, while
for trees of depth 2, it is equivalent to the Joint Replenishment Problem.
Aggregation problem for trees of arbitrary depth arise in multicasting, sensor
networks, communication in organization hierarchies, and in supply-chain
management. The instances of MLAP associated with these applications are
naturally online, in the sense that aggregation decisions need to be made
without information about future requests.
Constant-competitive online algorithms are known for MLAP with one or two
levels. However, it has been open whether there exist constant competitive
online algorithms for trees of depth more than 2. Addressing this open problem,
we give the first constant competitive online algorithm for networks of
arbitrary (fixed) number of levels. The competitive ratio is O(D^4 2^D), where
D is the depth of T. The algorithm works for arbitrary waiting cost functions,
including the variant with deadlines.
We also show several additional lower and upper bound results for some
special cases of MLAP, including the Single-Phase variant and the case when the
tree is a path
Online packet scheduling with bounded delay and lookahead
We study the online bounded-delay packet scheduling problem (Packet Scheduling), where packets of unit size arrive at a router over time and need to be transmitted over a network link. Each packet has two attributes: a non-negative weight and a deadline for its transmission. The objective is to maximize the total weight of the transmitted packets. This problem has been well studied in the literature; yet currently the best published upper bound is 1.828 [8],still quite far from the best lower bound ofφ≈1.618 [11, 2, 6].In the variant of Packet Scheduling with s-bounded instances, each packet can be scheduled in at most s consecutive slots, starting at its release time. The lower bound of φ applies even to the special case of 2-bounded instances, and a φ-competitive algorithm for 3-boundedinstances was given in [5]. Improving that result, and addressing a question posed by Goldwasser [9], we present a φ-competitive algorithm for 4-boundedinstances. We also study a variant of Packet Scheduling where an online algorithm has the additional power of1-lookahead, knowing at time t which packets will arrive at time t+ 1. For Packet Scheduling with 1-lookahead restricted to 2-bounded instances, we present an online algorithm with competitive ratio12(√13−1)≈1.303 and we prove a nearly tight lower boundof14(1 +√17)≈1.281. In fact, our lower bound result is more general: using only 2-boundedinstances, for any integer`≥0 we prove a lower bound of12(`+1)(1 +√5 + 8`+ 4`2) for online algorithms with`-look ahead, i.e., algorithms that at time t can see all packets arriving by time t+`. Finally, for non-restricted instances we show a lower bound of 1.25 for randomized algorithms with`-lookahead, for any`≥0
- …