1,381 research outputs found

    A Branch and Bound Approach to Optimal Allocation in Stratified Sampling

    Get PDF
    For practical applications of any allocations, integer values of the sample sizes are required. This could be done by simply rounding off the non-integer sample sizes to the nearest integral values. When the sample sizes are large enough or the measurement cost in various strata are not too high, the rounded off sample allocation may work well. However for small samples in some situations the rounding off allocations may become infeasible and non-optimal. This means that rounded off values may violate some of the constraints of the problem or there may exist other sets of integer sample allocations with a lesser value of the objective function. In such situations we have to use some integer programming technique to obtain an optimum integer solution. Keywords:  Stratified sampling, Non-linear Integer Programming, Allocation Problem,  Langrangian Multiplier, Branch & Bound Techniqu

    Estimation of Finite Population Mean in Multivariate Stratified Sampling under Cost Function Using Goal Programming

    Get PDF
    In practical utilization of stratified random sampling scheme, the investigator meets a problem to select a sample that maximizes the precision of a finite population mean under cost constraint. An allocation of sample size becomes complicated when more than one characteristic is observed from each selected unit in a sample. In many real life situations, a linear cost function of a sample size nh is not a good approximation to actual cost of sample survey when traveling cost between selected units in a stratum is significant. In this paper, sample allocation problem in multivariate stratified random sampling with proposed cost function is formulated in integer nonlinear multiobjective mathematical programming. A solution procedure is proposed using extended lexicographic goal programming approach. A numerical example is presented to illustrate the computational details and to compare the efficiency of proposed compromise allocation

    NEAR OPTIMUM MULTIVARIATE STRATIFIED SAMPLING DESIGN WITH RANDOM MEASUREMENT COSTS

    Get PDF
    Usually in sample surveys information on more than one characteristic are collected and the data obtained are analyzed to get the required estimates for the multivariate population under study. If stratified sampling design is to be applied on such a population the individual optimum allocations don’t help much unless the characteristics are highly correlated. Therefore, in multivariate stratified sampling we need to work out an allocation that is optimum for all characteristics in some sense, that is, near optimum for all characteristics. Such an allocation is called a compromise allocation. Furthermore, in surveys usually the per unit measurement costs are taken as deterministic, that is, they remain constant throughout the survey. In practice the costs of measurement of different characteristics in various strata may change during the course of survey for reasons beyond the control of the sampler. Thus in some practical situations the measurement costs may become a random variable and the  problem of obtaining a compromise allocation becomes a Stochastic Integer Nonlinear Programming Problem (SINLPP). The present paper addresses the problem of obtaining an integer compromise allocation for multivariate stratified sampling with random cost of measurements. A solution procedure has been developed for the formulated problem. A practical application of the procedure is also given through a numerical example to illustrate the computational details

    On The Efficiency of Some Techniques For Optimum Allocation In Multivariate Stratified Survey.

    Get PDF
    In multivariate stratified sampling, the major concern is on the problem of estimation of more than one population characteristics which often make conflicting demands on sampling technique. In this type of survey, an allocation which is optimum for one characteristic may not be optimum for other characteristics. In such situations a compromise criterion is needed to work out a usable allocation which is optimum for all characteristics in some sense. This study is focuses on the efficiency of some techniques for optimum sample allocation which are Yates/Chatterjee, Booth and Sedransk and Vector maximum criterion (VMC) on the set of real life data stratified into six strata and two variates with desired variances using: (i) method of minimum variance with fixed sample size and (ii) an arbitrary fixing of variances. The stratum sample sizes  among the classes were obtained to examine the criterion that will produce the smallest n. In this paper, it was discovered that VMC and Booth and Sedransk are superior to Yates/Chatterjee. Even though, no universal conclusion can be drawn, the work clearly brings out the fact that the best allocation is not always obvious and that sufficient care is necessary in the choice of allocation of the sample sizes to different strata with several items. Keywords:- Stratified Survey; Optimum Allocation; Vector Maximum Criterion (VMC); Yates/Chatterjee;Booth and Sedrans

    Stochastic-optimization of equipment productivity in multi-seam formations

    Get PDF
    Short and long range planning and execution for multi-seam coal formations (MSFs) are challenging with complex extraction mechanisms. Stripping equipment selection and scheduling are functions of the physical dynamics of the mine and the operational mechanisms of its components, thus its productivity is dependent on these parameters. Previous research studies did not incorporate quantitative relationships between equipment productivities and extraction dynamics in MSFs. The intrinsic variability of excavation and spoiling dynamics must also form part of existing models. This research formulates quantitative relationships of equipment productivities using Branch-and-Bound algorithms and Lagrange Parameterization approaches. The stochastic processes are resolved via Monte Carlo/Latin Hypercube simulation techniques within @RISK framework. The model was presented with a bituminous coal mining case in the Appalachian field. The simulated results showed a 3.51% improvement in mining cost and 0.19% increment in net present value. A 76.95ydÂł drop in productivity per unit change in cycle time was recorded for sub-optimal equipment schedules. The geologic variability and equipment operational parameters restricted any possible change in the cost function. A 50.3% chance of the mining cost increasing above its current value was driven by the volume of material re-handled with 0.52 regression coefficient. The study advances the optimization process in mine planning and scheduling algorithms, to efficiently capture future uncertainties surrounding multivariate random functions. The main novelty includes the application of stochastic-optimization procedures to improve equipment productivity in MSFs --Abstract, page iii

    Improving the representativeness of a simple random sample: an optimization model and its application to the continuous sample of working lives

    Get PDF
    This paper proposes an optimization model for selecting a larger subsample that improves the representativeness of a simple random sample previously obtained from a population larger than the population of interest. The problem formulation involves convex mixed-integer nonlinear programming (convex MINLP) and is, therefore, NP-hard. However, the solution is found by maximizing the size of the subsample taken from a stratified random sample with proportional allocation and restricting it to a p-value large enough to achieve a good fit to the population of interest using Pearson’s chi-square goodness-of-fit test. The paper also applies the model to the Continuous Sample of Working Lives (CSWL), which is a set of anonymized microdata containing information on individuals from Spanish Social Security records and the results prove that it is possible to obtain a larger subsample from the CSWL that (far) better represents the pensioner population for each of the waves analyzed

    Estimation of origin-destination matrices from traffic counts: theoretical and operational development

    Get PDF
    This thesis deals with the o-d estimation problem from indirect measures, addressing two main aspects of the problem: the identification of the set of indirect measures that provide the maximum information with a resulting reduction of the uncertainty on the estimate; once defined the set of measures, the choice of an estimator to identify univocally and as much reliable as possible the estimate. As regards the former aspect, an innovative and theoretically founded methodology is illustrated, explicitly accounting for the reliability of the o-d matrix estimate. The proposed approach is based on a specific measure, named Synthetic Dispersion Measure (SDM), related to the trace of the dispersion matrix of the posterior demand estimate conditioned to a given set of sensors locations. Under the mild assumption of multivariate normal distribution for the prior demand estimate, the proposed SDM does not depend on the specific values of the counted flows – unknown in the planning stage – but just on the locations of such sensors. The proposed approach is applied to real contexts, leading to results outperforming the other methods currently available in the literature. In addition, the proposed methodology allows setting a formal budget allocation problem between surveys and counts in the planning stage, in order to maximize the overall quality of the demand estimation process. As regard the latter aspect, a “quasi-dynamic” framework is proposed, under the assumption that o-d shares are constant across a reference period, whilst total flows leaving each origin vary for each sub-period within the reference period. The advantage of this approach over conventional within-day dynamic estimators is that of reducing drastically the number of unknowns given the same set of observed time-varying traffic counts. The quasi-dynamic assumption is checked by means of empirical and statistical tests and the performances of the quasi-dynamic estimator - whose formulation is also given – are compared with other dynamic estimators

    Matching on-the-fly in Sequential Experiments for Higher Power and Efficiency

    Full text link
    We propose a dynamic allocation procedure that increases power and efficiency when measuring an average treatment effect in sequential randomized trials. Subjects arrive iteratively and are either randomized or paired via a matching criterion to a previously randomized subject and administered the alternate treatment. We develop estimators for the average treatment effect that combine information from both the matched pairs and unmatched subjects as well as an exact test. Simulations illustrate the method's higher efficiency and power over competing allocation procedures in both controlled scenarios and historical experimental data.Comment: 20 pages, 1 algorithm, 2 figures, 8 table
    • …
    corecore