817,970 research outputs found

    Sampling-based Approximation Algorithms for Multi-stage Stochastic Optimization

    Get PDF
    Stochastic optimization problems provide a means to model uncertainty in the input data where the uncertainty is modeled by a probability distribution over the possible realizations of the data. We consider a broad class of these problems, called {it multi-stage stochastic programming problems with recourse}, where the uncertainty evolves through a series of stages and one take decisions in each stage in response to the new information learned. These problems are often computationally quite difficult with even very specialized (sub)problems being #P-complete. We obtain the first fully polynomial randomized approximation scheme (FPRAS) for a broad class of multi-stage stochastic linear programming problems with any constant number of stages, without placing any restrictions on the underlying probability distribution or on the cost structure of the input. For any fixed kk, for a rich class of kk-stage stochastic linear programs (LPs), we show that, for any probability distribution, for any epsilon>0epsilon>0, one can compute, with high probability, a solution with expected cost at most (1+e)(1+e) times the optimal expected cost, in time polynomial in the input size, frac1epsilonfrac{1}{epsilon}, and a parameter lambdalambda that is an upper bound on the cost-inflation over successive stages. Moreover, the algorithm analyzed is a simple and intuitive algorithm that is often used in practice, the {it sample average approximation} (SAA) method. In this method, one draws certain samples from the underlying distribution, constructs an approximate distribution from these samples, and solves the stochastic problem given by this approximate distribution. This is the first result establishing that the SAA method yields near-optimal solutions for (a class of) multi-stage programs with a polynomial number of samples. As a corollary of this FPRAS, by adapting a generic rounding technique of Shmoys and Swamy, we also obtain the first approximation algorithms for the analogous class of multi-stage stochastic integer programs, which includes the multi-stage versions of the set cover, vertex cover, multicut on trees, facility location, and multicommodity flow problems

    Assessing the effect of clustered and biased multi-stage sampling

    Get PDF
    We propose a method for detecting biased multi-stage sampling of spatial data and a method to adjust for biased Clustering of samples. We assess the effect of these methods for the analysis of radioactivity contamination data from Rongelap island, with the scientific problem being the estimation of the maximum level of radioactivity over the island. These data were collected over a two-stage process of uniform and clustered samples, which may have an impact on conclusions from a standard analysis that does not account for either of these features.info:eu-repo/semantics/publishedVersio

    Distilled Sensing: Adaptive Sampling for Sparse Detection and Estimation

    Full text link
    Adaptive sampling results in dramatic improvements in the recovery of sparse signals in white Gaussian noise. A sequential adaptive sampling-and-refinement procedure called Distilled Sensing (DS) is proposed and analyzed. DS is a form of multi-stage experimental design and testing. Because of the adaptive nature of the data collection, DS can detect and localize far weaker signals than possible from non-adaptive measurements. In particular, reliable detection and localization (support estimation) using non-adaptive samples is possible only if the signal amplitudes grow logarithmically with the problem dimension. Here it is shown that using adaptive sampling, reliable detection is possible provided the amplitude exceeds a constant, and localization is possible when the amplitude exceeds any arbitrarily slowly growing function of the dimension.Comment: 23 pages, 2 figures. Revision includes minor clarifications, along with more illustrative experimental results (cf. Figure 2

    Non-Uniform Time Sampling for Multiple-Frequency Harmonic Balance Computations

    Get PDF
    A time-domain harmonic balance method for the analysis of almost-periodic (multi-harmonics) flows is presented. This method relies on Fourier analysis to derive an efficient alternative to classical time marching schemes for such flows. It has recently received significant attention, especially in the turbomachinery field where the flow spectrum is essentially a combination of the blade passing frequencies. Up to now, harmonic balance methods have used a uniform time sampling of the period of interest, but in the case of several frequencies, non-necessarily multiple of each other, harmonic balance methods can face stability issues due to a bad condition number of the Fourier operator. Two algorithms are derived to find a non-uniform time sampling in order to minimize this condition number. Their behavior is studied on a wide range of frequencies, and a model problem of a 1D flow with pulsating outlet pressure, which enables to prove their efficiency. Finally, the flow in a multi-stage axial compressor is analyzed with different frequency sets. It demonstrates the stability and robustness of the present non-uniform harmonic balance method regardless of the frequency set

    Approximation with Error Bounds in Spark

    Full text link
    We introduce a sampling framework to support approximate computing with estimated error bounds in Spark. Our framework allows sampling to be performed at the beginning of a sequence of multiple transformations ending in an aggregation operation. The framework constructs a data provenance tree as the computation proceeds, then combines the tree with multi-stage sampling and population estimation theories to compute error bounds for the aggregation. When information about output keys are available early, the framework can also use adaptive stratified reservoir sampling to avoid (or reduce) key losses in the final output and to achieve more consistent error bounds across popular and rare keys. Finally, the framework includes an algorithm to dynamically choose sampling rates to meet user specified constraints on the CDF of error bounds in the outputs. We have implemented a prototype of our framework called ApproxSpark, and used it to implement five approximate applications from different domains. Evaluation results show that ApproxSpark can (a) significantly reduce execution time if users can tolerate small amounts of uncertainties and, in many cases, loss of rare keys, and (b) automatically find sampling rates to meet user specified constraints on error bounds. We also explore and discuss extensively trade-offs between sampling rates, execution time, accuracy and key loss
    corecore