260 research outputs found

    Limitations of Algebraic Approaches to Graph Isomorphism Testing

    Full text link
    We investigate the power of graph isomorphism algorithms based on algebraic reasoning techniques like Gr\"obner basis computation. The idea of these algorithms is to encode two graphs into a system of equations that are satisfiable if and only if if the graphs are isomorphic, and then to (try to) decide satisfiability of the system using, for example, the Gr\"obner basis algorithm. In some cases this can be done in polynomial time, in particular, if the equations admit a bounded degree refutation in an algebraic proof systems such as Nullstellensatz or polynomial calculus. We prove linear lower bounds on the polynomial calculus degree over all fields of characteristic different from 2 and also linear lower bounds for the degree of Positivstellensatz calculus derivations. We compare this approach to recently studied linear and semidefinite programming approaches to isomorphism testing, which are known to be related to the combinatorial Weisfeiler-Lehman algorithm. We exactly characterise the power of the Weisfeiler-Lehman algorithm in terms of an algebraic proof system that lies between degree-k Nullstellensatz and degree-k polynomial calculus

    Compact relaxations for polynomial programming problems

    Get PDF
    Reduced RLT constraints are a special class of Reformulation- Linearization Technique (RLT) constraints. They apply to nonconvex (both continuous and mixed-integer) quadratic programming problems subject to systems of linear equality constraints. We present an extension to the general case of polynomial programming problems and discuss the derived convex relaxation. We then show how to perform rRLT constraint generation so as to reduce the number of inequality constraints in the relaxation, thereby making it more compact and faster to solve. We present some computational results validating our approach

    Impact of cardiologist intervention on guideline-directed use of statin therapy

    Get PDF
    © The Author(s) 2020. Published by Baishideng Publishing Group Inc. All rights reserved. BACKGROUND Statins have an important and well-established role in the prevention of atherosclerotic cardiovascular disease (ASCVD). However, several studies have reported widespread underuse of statins in various practice settings and populations. Review of relevant literature reveals opportunities for improvement in the implementation of guideline-directed statin therapy (GDST). AIM To examine the impact of cardiologist intervention on the use of GDST in the ambulatory setting. METHODS Patients with at least one encounter at the adult Internal Medicine Clinic (IMC) and/or Cardiology Clinic (CC), who had an available serum cholesterol test performed, were evaluated. The 2 comparison groups were defined as: (1) Patients only seen by IMC; and (2) Patients seen by both IMC and CC. Patients were excluded if variables needed for calculation of ASCVD risk scores were lacking, and if demographic information lacked guideline-directed treatment recommendations. Data were analyzed using student t-tests or χ2, as appropriate. Analysis of Variance was used to compare rates of adherence to GDST. RESULTS A total of 268 patients met the inclusion criteria for this study; 211 in the IMC group and 57 in the IMC-CC group. Overall, 56% of patients were female, mean age 56 years (± 10.65, SD), 22% Black or African American, 56% Hispanic/Latino, 14% had clinical ASCVD, 13% current smokers, 66% diabetic and 63% hypertensive. Statin use was observed in 55% (n = 147/268) of the entire patient cohort. In the IMC-CC group, 73.6% (n = 42/57) of patients were prescribed statin therapy compared to 50.7% (n = 107/211) of patients in the IMC group (P = 0.002). In terms of appropriate statin use based on guidelines, there was no statistical difference between groups [IMC-CC group 61.4% (n = 35/57) vs IMC group, 55.5% (n = 117/211), P = 0.421]. Patients in the IMC-CC group were older, had more cardiac risk factors and had higher proportions of non-white patients compared to the IMC group (P \u3c 0.02, all). CONCLUSION Although overall use of GDST was suboptimal, there was no statistical difference in appropriate statin use based on guidelines between groups managed by general internists alone or co-managed with a cardiologist. These findings highlight the need to design and implement strategies to improve adherence rates to GDST across all specialties

    On the composition of convex envelopes for quadrilinear terms

    Get PDF
    International audienceWithin the framework of the spatial Branch-and-Bound algorithm for solving Mixed-Integer Nonlinear Programs, different convex relaxations can be obtained for multilinear terms by applying associativity in different ways. The two groupings ((x1x2)x3)x4 and (x1x2x3)x4 of a quadrilinear term, for example, give rise to two different convex relaxations. In [6] we prove that having fewer groupings of longer terms yields tighter convex relaxations. In this paper we give an alternative proof of the same fact and perform a computational study to assess the impact of the tightened convex relaxation in a spatial Branch-and-Bound setting

    An FPTAS for optimizing a class of low-rank functions over a polytope

    Get PDF
    We present a fully polynomial time approximation scheme (FPTAS) for optimizing a very general class of non-linear functions of low rank over a polytope. Our approximation scheme relies on constructing an approximate Pareto-optimal front of the linear functions which constitute the given low-rank function. In contrast to existing results in the literature, our approximation scheme does not require the assumption of quasi-concavity on the objective function. For the special case of quasi-concave function minimization, we give an alternative FPTAS, which always returns a solution which is an extreme point of the polytope. Our technique can also be used to obtain an FPTAS for combinatorial optimization problems with non-linear objective functions, for example when the objective is a product of a fixed number of linear functions. We also show that it is not possible to approximate the minimum of a general concave function over the unit hypercube to within any factor, unless P = NP. We prove this by showing a similar hardness of approximation result for supermodular function minimization, a result that may be of independent interest

    Biclustering via optimal re-ordering of data matrices in systems biology: rigorous methods and comparative studies

    Get PDF
    <p>Abstract</p> <p>Background</p> <p>The analysis of large-scale data sets via clustering techniques is utilized in a number of applications. Biclustering in particular has emerged as an important problem in the analysis of gene expression data since genes may only jointly respond over a subset of conditions. Biclustering algorithms also have important applications in sample classification where, for instance, tissue samples can be classified as cancerous or normal. Many of the methods for biclustering, and clustering algorithms in general, utilize simplified models or heuristic strategies for identifying the "best" grouping of elements according to some metric and cluster definition and thus result in suboptimal clusters.</p> <p>Results</p> <p>In this article, we present a rigorous approach to biclustering, OREO, which is based on the Optimal RE-Ordering of the rows and columns of a data matrix so as to globally minimize the dissimilarity metric. The physical permutations of the rows and columns of the data matrix can be modeled as either a network flow problem or a traveling salesman problem. Cluster boundaries in one dimension are used to partition and re-order the other dimensions of the corresponding submatrices to generate biclusters. The performance of OREO is tested on (a) metabolite concentration data, (b) an image reconstruction matrix, (c) synthetic data with implanted biclusters, and gene expression data for (d) colon cancer data, (e) breast cancer data, as well as (f) yeast segregant data to validate the ability of the proposed method and compare it to existing biclustering and clustering methods.</p> <p>Conclusion</p> <p>We demonstrate that this rigorous global optimization method for biclustering produces clusters with more insightful groupings of similar entities, such as genes or metabolites sharing common functions, than other clustering and biclustering algorithms and can reconstruct underlying fundamental patterns in the data for several distinct sets of data matrices arising in important biological applications.</p

    Efficient computation of the outer hull of a discrete path

    Get PDF
    We present here a linear time and space algorithm for computing the outer hull of any discrete path encoded by its Freeman chain code. The basic data structure uses an enriched version of the data structure introduced by Brlek, Koskas and Provençal: using quadtrees for representing points in the discrete plane ℤ×ℤ with neighborhood links, deciding path intersection is achievable in linear time and space. By combining the well-known wall follower algorithm for traversing mazes, we obtain the desired result with two passes resulting in a global linear time and space algorithm. As a byproduct, the convex hull is obtained as well

    Towards Machine Wald

    Get PDF
    The past century has seen a steady increase in the need of estimating and predicting complex systems and making (possibly critical) decisions with limited information. Although computers have made possible the numerical evaluation of sophisticated statistical models, these models are still designed \emph{by humans} because there is currently no known recipe or algorithm for dividing the design of a statistical model into a sequence of arithmetic operations. Indeed enabling computers to \emph{think} as \emph{humans} have the ability to do when faced with uncertainty is challenging in several major ways: (1) Finding optimal statistical models remains to be formulated as a well posed problem when information on the system of interest is incomplete and comes in the form of a complex combination of sample data, partial knowledge of constitutive relations and a limited description of the distribution of input random variables. (2) The space of admissible scenarios along with the space of relevant information, assumptions, and/or beliefs, tend to be infinite dimensional, whereas calculus on a computer is necessarily discrete and finite. With this purpose, this paper explores the foundations of a rigorous framework for the scientific computation of optimal statistical estimators/models and reviews their connections with Decision Theory, Machine Learning, Bayesian Inference, Stochastic Optimization, Robust Optimization, Optimal Uncertainty Quantification and Information Based Complexity.Comment: 37 page
    • …
    corecore