2,380 research outputs found

    Dependent randomized rounding for clustering and partition systems with knapsack constraints

    Full text link
    Clustering problems are fundamental to unsupervised learning. There is an increased emphasis on fairness in machine learning and AI; one representative notion of fairness is that no single demographic group should be over-represented among the cluster-centers. This, and much more general clustering problems, can be formulated with "knapsack" and "partition" constraints. We develop new randomized algorithms targeting such problems, and study two in particular: multi-knapsack median and multi-knapsack center. Our rounding algorithms give new approximation and pseudo-approximation algorithms for these problems. One key technical tool, which may be of independent interest, is a new tail bound analogous to Feige (2006) for sums of random variables with unbounded variances. Such bounds are very useful in inferring properties of large networks using few samples

    A Lagrangian relaxation approach to the edge-weighted clique problem

    Get PDF
    The bb-clique polytope CPbnCP^n_b is the convex hull of the node and edge incidence vectors of all subcliques of size at most bb of a complete graph on nn nodes. Including the Boolean quadric polytope QPnQP^n as a special case and being closely related to the quadratic knapsack polytope, it has received considerable attention in the literature. In particular, the max-cut problem is equivalent with optimizing a linear function over QPnnQP^n_n. The problem of optimizing linear functions over CPbnCP^n_b has so far been approached via heuristic combinatorial algorithms and cutting-plane methods. We study the structure of CPbnCP^n_b in further detail and present a new computational approach to the linear optimization problem based on Lucena's suggestion of integrating cutting planes into a Lagrangian relaxation of an integer programming problem. In particular, we show that the separation problem for tree inequalities becomes polynomial in our Lagrangian framework. Finally, computational results are presented. \u

    Sparse grid quadrature on products of spheres

    Full text link
    We examine sparse grid quadrature on weighted tensor products (WTP) of reproducing kernel Hilbert spaces on products of the unit sphere, in the case of worst case quadrature error for rules with arbitrary quadrature weights. We describe a dimension adaptive quadrature algorithm based on an algorithm of Hegland (2003), and also formulate a version of Wasilkowski and Wozniakowski's WTP algorithm (1999), here called the WW algorithm. We prove that the dimension adaptive algorithm is optimal in the sense of Dantzig (1957) and therefore no greater in cost than the WW algorithm. Both algorithms therefore have the optimal asymptotic rate of convergence given by Theorem 3 of Wasilkowski and Wozniakowski (1999). A numerical example shows that, even though the asymptotic convergence rate is optimal, if the dimension weights decay slowly enough, and the dimensionality of the problem is large enough, the initial convergence of the dimension adaptive algorithm can be slow.Comment: 34 pages, 6 figures. Accepted 7 January 2015 for publication in Numerical Algorithms. Revised at page proof stage to (1) update email address; (2) correct the accent on "Wozniakowski" on p. 7; (3) update reference 2; (4) correct references 3, 18 and 2
    • …
    corecore