26,402 research outputs found
Coarse-grained complexity for dynamic algorithms
To date, the only way to argue polynomial lower bounds for dynamic algorithms is via fine-grained complexity arguments. These arguments rely on strong assumptions about specific problems such as the Strong Exponential Time Hypothesis (SETH) and the Online Matrix-Vector Multiplication Conjecture (OMv). While they have led to many exciting discoveries, dynamic algorithms still miss out some benefits and lessons from the traditional âcoarse-grainedâ approach that relates together classes of problems such as P and NP. In this paper we initiate the study of coarse-grained complexity theory for dynamic algorithms. Below are among questions that this theory can answer
Fine Grained Component Engineering of Adaptive Overlays: Experiences and Perspectives
Recent years have seen significant research being carried out into peer-to-peer (P2P) systems. This work has focused on the styles and applications of P2P computing, from grid computation to content distribution; however, little investigation has been performed into how these systems are built. Component based engineering is an approach that has seen successful deployment in the field of middleware development; functionality is encapsulated in âbuilding blocksâ that can be dynamically plugged together to form complete systems. This allows efficient, flexible and adaptable systems to be built with lower overhead and development complexity. This paper presents an investigation into the potential of using component based engineering in the design and construction of peer-to-peer overlays. It is highlighted that the quality of these properties is dictated by the component architecture used to implement the system. Three reusable decomposition architectures are designed and evaluated using Chord and Pastry case studies. These demonstrate that significant improvements can be made over traditional design approaches resulting in much more reusable, (re)configurable and extensible systems
A survey of parallel algorithms for fractal image compression
This paper presents a short survey of the key research work that has been undertaken in the application of parallel algorithms for Fractal image compression. The interest in fractal image compression techniques stems from their ability to achieve high compression ratios whilst maintaining a very high quality in the reconstructed image. The main drawback of this compression method is the very high computational cost that is associated with the encoding phase. Consequently, there has been significant interest in exploiting parallel computing architectures in order to speed up this phase, whilst still maintaining the advantageous features of the approach. This paper presents a brief introduction to fractal image compression, including the iterated function system theory upon
which it is based, and then reviews the different techniques that have been, and can be, applied in order to parallelize the compression algorithm
Spatial multi-level interacting particle simulations and information theory-based error quantification
We propose a hierarchy of multi-level kinetic Monte Carlo methods for
sampling high-dimensional, stochastic lattice particle dynamics with complex
interactions. The method is based on the efficient coupling of different
spatial resolution levels, taking advantage of the low sampling cost in a
coarse space and by developing local reconstruction strategies from
coarse-grained dynamics. Microscopic reconstruction corrects possibly
significant errors introduced through coarse-graining, leading to the
controlled-error approximation of the sampled stochastic process. In this
manner, the proposed multi-level algorithm overcomes known shortcomings of
coarse-graining of particle systems with complex interactions such as combined
long and short-range particle interactions and/or complex lattice geometries.
Specifically, we provide error analysis for the approximation of long-time
stationary dynamics in terms of relative entropy and prove that information
loss in the multi-level methods is growing linearly in time, which in turn
implies that an appropriate observable in the stationary regime is the
information loss of the path measures per unit time. We show that this
observable can be either estimated a priori, or it can be tracked
computationally a posteriori in the course of a simulation. The stationary
regime is of critical importance to molecular simulations as it is relevant to
long-time sampling, obtaining phase diagrams and in studying metastability
properties of high-dimensional complex systems. Finally, the multi-level nature
of the method provides flexibility in combining rejection-free and null-event
implementations, generating a hierarchy of algorithms with an adjustable number
of rejections that includes well-known rejection-free and null-event
algorithms.Comment: 34 page
- âŠ