45 research outputs found
Cut-Matching Games on Directed Graphs
We give O(log^2 n)-approximation algorithm based on the cut-matching
framework of [10, 13, 14] for computing the sparsest cut on directed graphs.
Our algorithm uses only O(log^2 n) single commodity max-flow computations and
thus breaks the multicommodity-flow barrier for computing the sparsest cut on
directed graph
Image reconstruction in optical interferometry: Benchmarking the regularization
With the advent of infrared long-baseline interferometers with more than two
telescopes, both the size and the completeness of interferometric data sets
have significantly increased, allowing images based on models with no a priori
assumptions to be reconstructed. Our main objective is to analyze the multiple
parameters of the image reconstruction process with particular attention to the
regularization term and the study of their behavior in different situations.
The secondary goal is to derive practical rules for the users. Using the
Multi-aperture image Reconstruction Algorithm (MiRA), we performed multiple
systematic tests, analyzing 11 regularization terms commonly used. The tests
are made on different astrophysical objects, different (u,v) plane coverages
and several signal-to-noise ratios to determine the minimal configuration
needed to reconstruct an image. We establish a methodology and we introduce the
mean-square errors (MSE) to discuss the results. From the ~24000 simulations
performed for the benchmarking of image reconstruction with MiRA, we are able
to classify the different regularizations in the context of the observations.
We find typical values of the regularization weight. A minimal (u,v) coverage
is required to reconstruct an acceptable image, whereas no limits are found for
the studied values of the signal-to-noise ratio. We also show that
super-resolution can be achieved with increasing performance with the (u,v)
coverage filling. Using image reconstruction with a sufficient (u,v) coverage
is shown to be reliable. The choice of the main parameters of the
reconstruction is tightly constrained. We recommend that efforts to develop
interferometric infrastructures should first concentrate on the number of
telescopes to combine, and secondly on improving the accuracy and sensitivity
of the arrays.Comment: 15 pages, 16 figures; accepted in A&
Graph Clustering in All Parameter Regimes
Resolution parameters in graph clustering control the size and structure of clusters formed by solving a parametric objective function. Typically there is more than one meaningful way to cluster a graph, and solving the same objective function for different resolution parameters produces clusterings at different levels of granularity, each of which can be meaningful depending on the application. In this paper, we address the task of efficiently solving a parameterized graph clustering objective for all values of a resolution parameter. Specifically, we consider a new analysis-friendly objective we call LambdaPrime, involving a parameter ? ? (0,1). LambdaPrime is an adaptation of LambdaCC, a significant family of instances of the Correlation Clustering (minimization) problem. Indeed, LambdaPrime and LambdaCC are closely related to other parameterized clustering problems, such as parametric generalizations of modularity. They capture a number of specific clustering problems as special cases, including sparsest cut and cluster deletion. While previous work provides approximation results for a single value of the resolution parameter, we seek a set of approximately optimal clusterings for all values of ? in polynomial time.
More specifically, we show that when a graph has m edges and n nodes, there exists a set of at most m clusterings such that, for every ? ? (0,1), the family contains an optimal solution to the LambdaPrime objective. This bound is tight on star graphs. We obtain a family of O(log n) clusterings by solving the parametric linear programming (LP) relaxation of LambdaPrime at O(log n) ? values, and rounding each LP solution using existing approximation algorithms. We prove that this is asymptotically tight: for a certain class of ring graphs, for all values of ?, ?(log n) feasible solutions are required to provide a constant-factor approximation for the LambdaPrime LP relaxation. To minimize the size of the clustering family, we further propose an algorithm that yields a family of solutions of a size no more than twice of the minimum LP-approximating family
Quantum speedup for graph sparsification, cut approximation, and Laplacian solving
Graph sparsification underlies a large number of algorithms, ranging from approximation algorithms for cut problems to solvers for linear systems in the graph Laplacian. In its strongest form, “spectral sparsification” reduces the number of edges to near-linear in the number of nodes, while approximately preserving the cut and spectral structure of the graph. In this work we demonstrate a polynomial quantum speedup for spectral sparsification and many of its applications. In particular, we give a quantum algorithm that, given a weighted graph with n nodes and m edges, outputs a classical description of an ϵ -spectral sparsifier in sublinear time O˜(mn−−−√/ϵ) . This contrasts with the optimal classical complexity O˜(m) . We also prove that our quantum algorithm is optimal up to polylog-factors. The algorithm builds on a string of existing results on sparsification, graph spanners, quantum algorithms for shortest paths, and efficient constructions for k -wise independent random strings. Our algorithm implies a quantum speedup for solving Laplacian systems and for approximating a range of cut problems such as min cut and sparsest cut
Fitting Prediction Rule Ensembles with R Package pre
Prediction rule ensembles (PREs) are sparse collections of rules, offering
highly interpretable regression and classification models. This paper presents
the R package pre, which derives PREs through the methodology of Friedman and
Popescu (2008). The implementation and functionality of package pre is
described and illustrated through application on a dataset on the prediction of
depression. Furthermore, accuracy and sparsity of PREs is compared with that of
single trees, random forest and lasso regression in four benchmark datasets.
Results indicate that pre derives ensembles with predictive accuracy comparable
to that of random forests, while using a smaller number of variables for
prediction
A Simple Framework for Finding Balanced Sparse Cuts via APSP
We present a very simple and intuitive algorithm to find balanced sparse cuts
in a graph via shortest-paths. Our algorithm combines a new
multiplicative-weights framework for solving unit-weight multi-commodity flows
with standard ball growing arguments. Using Dijkstra's algorithm for computing
the shortest paths afresh every time gives a very simple algorithm that runs in
time and finds an -sparse
balanced cut, when the given graph has a -sparse balanced cut. Combining
our algorithm with known deterministic data-structures for answering
approximate All Pairs Shortest Paths (APSP) queries under increasing edge
weights (decremental setting), we obtain a simple deterministic algorithm that
finds -sparse balanced cuts in time. Our
deterministic almost-linear time algorithm matches the state-of-the-art in
randomized and deterministic settings up to subpolynomial factors, while being
significantly simpler to understand and analyze, especially compared to the
only almost-linear time deterministic algorithm, a recent breakthrough by
Chuzhoy-Gao-Li-Nanongkai-Peng-Saranurak (FOCS 2020)