817,970 research outputs found
Sampling-based Approximation Algorithms for Multi-stage Stochastic Optimization
Stochastic optimization problems provide a means to model uncertainty in the input data where the uncertainty is modeled by a probability distribution over the possible realizations of the data. We consider a broad class of these problems, called {it multi-stage stochastic programming problems with recourse}, where the uncertainty evolves through a series of stages and one take decisions in each stage in response to the new information learned. These problems are often computationally quite difficult with even very specialized (sub)problems being #P-complete.
We obtain the first fully polynomial randomized approximation scheme (FPRAS) for a broad class of multi-stage stochastic linear programming problems with any constant number of stages, without placing any restrictions on the underlying probability distribution or on the cost structure of the input. For any fixed , for a rich class of -stage stochastic linear programs (LPs), we show that, for any probability distribution, for any , one can compute, with high probability, a solution with expected cost at most times the optimal expected cost, in time polynomial in the input size, , and a parameter that is an upper bound on the cost-inflation over successive stages. Moreover, the algorithm analyzed is a simple and intuitive algorithm that is often used in practice, the {it sample average approximation} (SAA) method. In this method, one draws certain samples from the underlying distribution, constructs an approximate distribution from these samples, and solves the stochastic problem given by this approximate distribution. This is the first result establishing that the SAA method yields near-optimal solutions for (a class of) multi-stage programs with a polynomial number of samples.
As a corollary of this FPRAS, by adapting a generic rounding technique of Shmoys and Swamy, we also obtain the first approximation algorithms for the analogous class of multi-stage stochastic integer programs, which includes the multi-stage versions of the set cover, vertex cover, multicut on trees, facility location, and multicommodity flow problems
Assessing the effect of clustered and biased multi-stage sampling
We propose a method for detecting biased multi-stage sampling of spatial data and a method to adjust for biased Clustering of samples. We assess the effect of these methods for the analysis of radioactivity contamination data from Rongelap island, with the scientific problem being the estimation of the maximum level of radioactivity over the island. These data were collected over a two-stage process of uniform and clustered samples, which may have an impact on conclusions from a standard analysis that does not account for either of these features.info:eu-repo/semantics/publishedVersio
Distilled Sensing: Adaptive Sampling for Sparse Detection and Estimation
Adaptive sampling results in dramatic improvements in the recovery of sparse
signals in white Gaussian noise. A sequential adaptive sampling-and-refinement
procedure called Distilled Sensing (DS) is proposed and analyzed. DS is a form
of multi-stage experimental design and testing. Because of the adaptive nature
of the data collection, DS can detect and localize far weaker signals than
possible from non-adaptive measurements. In particular, reliable detection and
localization (support estimation) using non-adaptive samples is possible only
if the signal amplitudes grow logarithmically with the problem dimension. Here
it is shown that using adaptive sampling, reliable detection is possible
provided the amplitude exceeds a constant, and localization is possible when
the amplitude exceeds any arbitrarily slowly growing function of the dimension.Comment: 23 pages, 2 figures. Revision includes minor clarifications, along
with more illustrative experimental results (cf. Figure 2
Non-Uniform Time Sampling for Multiple-Frequency Harmonic Balance Computations
A time-domain harmonic balance method for the analysis of almost-periodic (multi-harmonics) flows is presented. This method relies on Fourier analysis to derive an efficient alternative to classical time marching schemes for such flows. It has recently received significant attention, especially in the turbomachinery field where the flow spectrum is essentially a combination of the blade passing frequencies. Up to now, harmonic balance methods have used a uniform time sampling of the period of interest, but in the case of several frequencies, non-necessarily multiple of each other, harmonic balance methods can face stability issues due to a bad condition number of the Fourier operator. Two algorithms are derived to find a non-uniform time sampling in order to minimize this condition number. Their behavior is studied on a wide range of frequencies, and a model problem of a 1D flow with pulsating outlet pressure, which enables to prove their efficiency. Finally, the flow in a multi-stage axial compressor is analyzed with different frequency sets. It demonstrates the stability and robustness of the present non-uniform harmonic balance method regardless of the frequency set
Approximation with Error Bounds in Spark
We introduce a sampling framework to support approximate computing with
estimated error bounds in Spark. Our framework allows sampling to be performed
at the beginning of a sequence of multiple transformations ending in an
aggregation operation. The framework constructs a data provenance tree as the
computation proceeds, then combines the tree with multi-stage sampling and
population estimation theories to compute error bounds for the aggregation.
When information about output keys are available early, the framework can also
use adaptive stratified reservoir sampling to avoid (or reduce) key losses in
the final output and to achieve more consistent error bounds across popular and
rare keys. Finally, the framework includes an algorithm to dynamically choose
sampling rates to meet user specified constraints on the CDF of error bounds in
the outputs. We have implemented a prototype of our framework called
ApproxSpark, and used it to implement five approximate applications from
different domains. Evaluation results show that ApproxSpark can (a)
significantly reduce execution time if users can tolerate small amounts of
uncertainties and, in many cases, loss of rare keys, and (b) automatically find
sampling rates to meet user specified constraints on error bounds. We also
explore and discuss extensively trade-offs between sampling rates, execution
time, accuracy and key loss
- …