72 research outputs found

    Boolean Operations, Joins, and the Extended Low Hierarchy

    Get PDF
    We prove that the join of two sets may actually fall into a lower level of the extended low hierarchy than either of the sets. In particular, there exist sets that are not in the second level of the extended low hierarchy, EL_2, yet their join is in EL_2. That is, in terms of extended lowness, the join operator can lower complexity. Since in a strong intuitive sense the join does not lower complexity, our result suggests that the extended low hierarchy is unnatural as a complexity measure. We also study the closure properties of EL_ and prove that EL_2 is not closed under certain Boolean operations. To this end, we establish the first known (and optimal) EL_2 lower bounds for certain notions generalizing Selman's P-selectivity, which may be regarded as an interesting result in its own right.Comment: 12 page

    Downward Collapse from a Weaker Hypothesis

    Full text link
    Hemaspaandra et al. proved that, for m>0m > 0 and 0<i<kβˆ’10 < i < k - 1: if \Sigma_i^p \BoldfaceDelta DIFF_m(\Sigma_k^p) is closed under complementation, then DIFFm(Ξ£kp)=coDIFFm(Ξ£kp)DIFF_m(\Sigma_k^p) = coDIFF_m(\Sigma_k^p). This sharply asymmetric result fails to apply to the case in which the hypothesis is weakened by allowing the Ξ£ip\Sigma_i^p to be replaced by any class in its difference hierarchy. We so extend the result by proving that, for s,m>0s,m > 0 and 0<i<kβˆ’10 < i < k - 1: if DIFF_s(\Sigma_i^p) \BoldfaceDelta DIFF_m(\Sigma_k^p) is closed under complementation, then DIFFm(Ξ£kp)=coDIFFm(Ξ£kp)DIFF_m(\Sigma_k^p) = coDIFF_m(\Sigma_k^p)

    Robust optimization with incremental recourse

    Full text link
    In this paper, we consider an adaptive approach to address optimization problems with uncertain cost parameters. Here, the decision maker selects an initial decision, observes the realization of the uncertain cost parameters, and then is permitted to modify the initial decision. We treat the uncertainty using the framework of robust optimization in which uncertain parameters lie within a given set. The decision maker optimizes so as to develop the best cost guarantee in terms of the worst-case analysis. The recourse decision is ``incremental"; that is, the decision maker is permitted to change the initial solution by a small fixed amount. We refer to the resulting problem as the robust incremental problem. We study robust incremental variants of several optimization problems. We show that the robust incremental counterpart of a linear program is itself a linear program if the uncertainty set is polyhedral. Hence, it is solvable in polynomial time. We establish the NP-hardness for robust incremental linear programming for the case of a discrete uncertainty set. We show that the robust incremental shortest path problem is NP-complete when costs are chosen from a polyhedral uncertainty set, even in the case that only one new arc may be added to the initial path. We also address the complexity of several special cases of the robust incremental shortest path problem and the robust incremental minimum spanning tree problem
    • …
    corecore