6 research outputs found

    The Limitations of Optimization from Samples

    Full text link
    In this paper we consider the following question: can we optimize objective functions from the training data we use to learn them? We formalize this question through a novel framework we call optimization from samples (OPS). In OPS, we are given sampled values of a function drawn from some distribution and the objective is to optimize the function under some constraint. While there are interesting classes of functions that can be optimized from samples, our main result is an impossibility. We show that there are classes of functions which are statistically learnable and optimizable, but for which no reasonable approximation for optimization from samples is achievable. In particular, our main result shows that there is no constant factor approximation for maximizing coverage functions under a cardinality constraint using polynomially-many samples drawn from any distribution. We also show tight approximation guarantees for maximization under a cardinality constraint of several interesting classes of functions including unit-demand, additive, and general monotone submodular functions, as well as a constant factor approximation for monotone submodular functions with bounded curvature

    Testing Submodularity

    Get PDF
    We show that for any constants ϵ>0\epsilon > 0 and p≥1p \ge 1, given oracle access to an unknown function f:{0,1}n→[0,1]f : \{0,1\}^n \to [0,1] it is possible to determine if the function is submodular or is ϵ\epsilon-far from every submodular function, in ℓp\ell_p distance, with a \emph{constant} number of queries to the oracle. We refer to the process of determining if an unknown function has a property, or is far from every function having the property, as \emph{property testing}, and we refer to the algorithm that does that as a tester or a testing algorithm. A function f:{0,1}n→[0,1]f : \{0,1\}^n \to [0,1] is a \emph{kk-junta} if there is a set J⊆[n]J \subseteq [n] of cardinality ∣J∣≤k|J| \le k such that the value of ff on any input xx is completely determined by the values xix_i for i∈Ji \in J. For any constant ϵ>0\epsilon > 0 and a set of kk-juntas F\mathcal{F}, we give an algorithm which determines if an unknown function f:{0,1}n→[0,1]f : \{0,1\}^n \to [0,1] is ϵ106\frac{\epsilon}{10^6}-close to some function in F\mathcal{F} or is ϵ\epsilon-far from every function in F\mathcal{F}, in ℓ2\ell_2 distance, with a constant number of queries to the unknown function. This result, combined with a recent junta theorem of Feldman and \Vondrak (2016) in which they show every submodular function is ϵ\epsilon-close, in ℓ2\ell_2 distance, to another submodular function which is a O~(1ϵ2)\tilde{O}(\frac{1}{\epsilon^2})-junta, yields the constant-query testing algorithm for submodular functions. We also give constant-query testing algorithms for a variety of other natural properties of valuation functions, including fractionally additive (XOS) functions, OXS functions, unit demand functions, coverage functions, and self-bounding functions
    corecore