52,165 research outputs found

    Sample average approximation with heavier tails II: localization in stochastic convex optimization and persistence results for the Lasso

    Full text link
    We present exponential finite-sample nonasymptotic deviation inequalities for the SAA estimator's near-optimal solution set over the class of stochastic optimization problems with heavy-tailed random \emph{convex} functions in the objective and constraints. Such setting is better suited for problems where a sub-Gaussian data generating distribution is less expected, e.g., in stochastic portfolio optimization. One of our contributions is to exploit \emph{convexity} of the perturbed objective and the perturbed constraints as a property which entails \emph{localized} deviation inequalities for joint feasibility and optimality guarantees. This means that our bounds are significantly tighter in terms of diameter and metric entropy since they depend only on the near-optimal solution set but not on the whole feasible set. As a result, we obtain a much sharper sample complexity estimate when compared to a general nonconvex problem. In our analysis, we derive some localized deterministic perturbation error bounds for convex optimization problems which are of independent interest. To obtain our results, we only assume a metric regular convex feasible set, possibly not satisfying the Slater condition and not having a metric regular solution set. In this general setting, joint near feasibility and near optimality are guaranteed. If in addition the set satisfies the Slater condition, we obtain finite-sample simultaneous \emph{exact} feasibility and near optimality guarantees (for a sufficiently small tolerance). Another contribution of our work is to present, as a proof of concept of our localized techniques, a persistent result for a variant of the LASSO estimator under very weak assumptions on the data generating distribution.Comment: 34 pages. Some correction

    Width and extremal height distributions of fluctuating interfaces with window boundary conditions

    Full text link
    We present a detailed study of squared local roughness (SLRDs) and local extremal height distributions (LEHDs), calculated in windows of lateral size ll, for interfaces in several universality classes, in substrate dimensions ds=1d_s = 1 and ds=2d_s = 2. We show that their cumulants follow a Family-Vicsek type scaling, and, at early times, when ξl\xi \ll l (ξ\xi is the correlation length), the rescaled SLRDs are given by log-normal distributions, with their nnth cumulant scaling as (ξ/l)(n1)ds(\xi/l)^{(n-1)d_s}. This give rise to an interesting temporal scaling for such cumulants wnctγn\left\langle w_n \right\rangle_c \sim t^{\gamma_n}, with γn=2nβ+(n1)ds/z=[2n+(n1)ds/α]β\gamma_n = 2 n \beta + {(n-1)d_s}/{z} = \left[ 2 n + {(n-1)d_s}/{\alpha} \right] \beta. This scaling is analytically proved for the Edwards-Wilkinson (EW) and Random Deposition interfaces, and numerically confirmed for other classes. In general, it is featured by small corrections and, thus, it yields exponents γn\gamma_n's (and, consequently, α\alpha, β\beta and zz) in nice agreement with their respective universality class. Thus, it is an useful framework for numerical and experimental investigations, where it is, usually, hard to estimate the dynamic zz and mainly the (global) roughness α\alpha exponents. The stationary (for ξl\xi \gg l) SLRDs and LEHDs of Kardar-Parisi-Zhang (KPZ) class are also investigated and, for some models, strong finite-size corrections are found. However, we demonstrate that good evidences of their universality can be obtained through successive extrapolations of their cumulant ratios for long times and large ll's. We also show that SLRDs and LEHDs are the same for flat and curved KPZ interfaces.Comment: 11 pages, 10 figures, 4 table

    Necessary Conditions for Extended Noncontextuality in General Sets of Random Variables

    Full text link
    We explore the graph approach to contextuality to restate the extended definition of noncontextuality as given by J. Kujala et. al. [Phys. Rev. Lett. 115, 150401 (2015)] using graph-theoretical terms. This extended definition avoids the assumption of the pre-sheaf or non-disturbance condition, which states that if two contexts overlap, then the marginal distribution obtained for the intersection must be the same, a restriction that will never be perfectly satisfied in real experiments. With this we are able to derive necessary conditions for extended noncontextuality for any set of random variables based on the geometrical aspects of the graph approach, which can be tested directly with experimental data in any contextuality experiment and which reduce to traditional necessary conditions for noncontextuality if the non-disturbance condition is satisfied

    Estimating graph parameters with random walks

    Full text link
    An algorithm observes the trajectories of random walks over an unknown graph GG, starting from the same vertex xx, as well as the degrees along the trajectories. For all finite connected graphs, one can estimate the number of edges mm up to a bounded factor in O(trel3/4m/d)O\left(t_{\mathrm{rel}}^{3/4}\sqrt{m/d}\right) steps, where trelt_{\mathrm{rel}} is the relaxation time of the lazy random walk on GG and dd is the minimum degree in GG. Alternatively, mm can be estimated in O(tunif+trel5/6n)O\left(t_{\mathrm{unif}} +t_{\mathrm{rel}}^{5/6}\sqrt{n}\right), where nn is the number of vertices and tunift_{\mathrm{unif}} is the uniform mixing time on GG. The number of vertices nn can then be estimated up to a bounded factor in an additional O(tunifmn)O\left(t_{\mathrm{unif}}\frac{m}{n}\right) steps. Our algorithms are based on counting the number of intersections of random walk paths X,YX,Y, i.e. the number of pairs (t,s)(t,s) such that Xt=YsX_t=Y_s. This improves on previous estimates which only consider collisions (i.e., times tt with Xt=YtX_t=Y_t). We also show that the complexity of our algorithms is optimal, even when restricting to graphs with a prescribed relaxation time. Finally, we show that, given either mm or the mixing time of GG, we can compute the "other parameter" with a self-stopping algorithm
    corecore