43 research outputs found
Complexity results for three-dimensional orthogonal graph drawing
AbstractIn this paper we consider the problem of finding three-dimensional orthogonal drawings of maximum degree six graphs from the computational complexity perspective. We introduce a 3SAT reduction framework that can be used to prove the NP-hardness of finding three-dimensional orthogonal drawings with specific constraints. By using the framework we show that, given a three-dimensional orthogonal shape of a graph (a description of the sequence of axis-parallel segments of each edge), finding the coordinates for nodes and bends such that the drawing has no intersection is NP-complete. Conversely, we show that if node coordinates are fixed, finding a shape for the edges that is compatible with a non-intersecting drawing is a feasible problem, which becomes NP-complete if a maximum of two bends per edge is allowed. We comment on the impact of these results on the two open problems of determining whether a graph always admits a drawing with at most two bends per edge and of characterizing orthogonal shapes admitting an orthogonal drawing without intersections
A nonmonotone GRASP
A greedy randomized adaptive search procedure (GRASP) is an itera-
tive multistart metaheuristic for difficult combinatorial optimization problems. Each
GRASP iteration consists of two phases: a construction phase, in which a feasible
solution is produced, and a local search phase, in which a local optimum in the
neighborhood of the constructed solution is sought. Repeated applications of the con-
struction procedure yields different starting solutions for the local search and the
best overall solution is kept as the result. The GRASP local search applies iterative
improvement until a locally optimal solution is found. During this phase, starting from
the current solution an improving neighbor solution is accepted and considered as the
new current solution. In this paper, we propose a variant of the GRASP framework that
uses a new “nonmonotone” strategy to explore the neighborhood of the current solu-
tion. We formally state the convergence of the nonmonotone local search to a locally
optimal solution and illustrate the effectiveness of the resulting Nonmonotone GRASP
on three classical hard combinatorial optimization problems: the maximum cut prob-
lem (MAX-CUT), the weighted maximum satisfiability problem (MAX-SAT), and
the quadratic assignment problem (QAP)
Complexity dichotomies for approximations of counting problems
Αυτή η διπλωματική αποτελεί μια επισκόπηση θεωρημάτων διχοτομίας για
υπολογιστικά προβλήματα, και ειδικότερα προβλήματα μέτρησης. Θεώρημα διχοτομίας
στην υπολογιστική πολυπλοκότητα είναι ένας πλήρης χαρασκτηρισμός των μελών μιας
κλάσης προβλημάτων, σε υπολογιστικά δύσκολα και υπολογιστικά εύκολα, χωρίς να
υπάρχουν προβλήματα ενδιάμεσης πολυπλοκότητας στην κλάση αυτή. Λόγω του
θεωρήματος του Ladner, δεν μπορούμε να έχουμε διχοτομία για ολόκληρες τις
κλάσεις NP και #P, παρόλα αυτά υπάρχουν μεγάλες υποκλάσεις της NP (#P) για τις
οποίες ισχύουν θεωρήματα διχοτομίας.
Συνεχίζουμε με την εκδοχή απόφασης του προβλήματος ικανοποίησης περιορισμών
(CSP), μία κλάση προβλήμάτων της NP στην οποία δεν εφαρμόζεται το θεώρημα του
Ladner. Δείχνουμε τα θεωρήματα διχοτομίας που υπάρχουν για ειδικές περιπτώσεις
του CSP. Στη συνέχεια επικεντρωνόμαστε στα προβλήματα μέτρησης παρουσιάζοντας
τα παρακάτω μοντέλα: Ομομορφισμοί γράφων, μετρητικό πρόβλημα ικανοποίησης
περιορισμών (#CSP), και προβλήματα Holant. Αναφέρουμε τα θεωρήματα διχοτομίας
που γνωρίζουμε γι' αυτά.
Στο τελευταίο και κύριο κεφάλαιο, χαλαρώνουμε την απαίτηση ακριβών υπολογισμών,
και αρκούμαστε στην προσέγγιση των προβλημάτων. Παρουσιάζουμε τα μέχρι σήμερα
γνωστά θεωρήματα κατάταξης για το #CSP. Πολλά ερωτήματα στην περιοχή παραμένουν
ανοιχτά.
Το παράρτημα είναι μια εισαγωγή στους ολογραφικούς αλγορίθμους, μία πρόσφατη
αλγοριθμική τεχνική για την εύρεση πολυωνυμικών αλγορίθμων (ακριβείς
υπολογισμοί) σε προβλήματα μέτρησης.This thesis is a survey of dichotomy theorems for computational problems,
focusing in counting problems. A dichotomy theorem in computational
complexity, is a complete classification of the members of a class of problems,
in computationally easy and computationally hard, with the set of problems of
intermediate
complexity being empty. Due to Ladner's theorem we cannot find a dichotomy
theorem for the whole classes NP and #P, however there are large subclasses of
NP (#P),
that model many "natural" problems, for which dichotomy theorems exist.
We continue with the decision version of constraint satisfaction problems
(CSP), a class of problems in NP, for which Ladner's theorem doesn't apply. We
obtain a
dichotomy theorem for some special cases of CSP. We then focus on counting
problems presenting the following frameworks: graph homomorphisms, counting
constraint
satisfaction (#CSP) and Holant problems; we provide the known dichotomies for
these frameworks.
In the last and main chapter of this thesis we relax the requirement of exact
computation, and settle in approximating the problems. We present the known
cassification theorems
for cases of #CSP. Many questions in terms of approximate counting problems
remain open.
The appendix introduces a recent technique for obtaining exact polynomial-time
algorithms for counting problems, namely the holographic algorithms
IST Austria Thesis
This dissertation focuses on algorithmic aspects of program verification, and presents modeling and complexity advances on several problems related to the
static analysis of programs, the stateless model checking of concurrent programs, and the competitive analysis of real-time scheduling algorithms.
Our contributions can be broadly grouped into five categories.
Our first contribution is a set of new algorithms and data structures for the quantitative and data-flow analysis of programs, based on the graph-theoretic notion of treewidth.
It has been observed that the control-flow graphs of typical programs have special structure, and are characterized as graphs of small treewidth.
We utilize this structural property to provide faster algorithms for the quantitative and data-flow analysis of recursive and concurrent programs.
In most cases we make an algebraic treatment of the considered problem,
where several interesting analyses, such as the reachability, shortest path, and certain kind of data-flow analysis problems follow as special cases.
We exploit the constant-treewidth property to obtain algorithmic improvements for on-demand versions of the problems,
and provide data structures with various tradeoffs between the resources spent in the preprocessing and querying phase.
We also improve on the algorithmic complexity of quantitative problems outside the algebraic path framework,
namely of the minimum mean-payoff, minimum ratio, and minimum initial credit for energy problems.
Our second contribution is a set of algorithms for Dyck reachability with applications to data-dependence analysis and alias analysis.
In particular, we develop an optimal algorithm for Dyck reachability on bidirected graphs, which are ubiquitous in context-insensitive, field-sensitive points-to analysis.
Additionally, we develop an efficient algorithm for context-sensitive data-dependence analysis via Dyck reachability,
where the task is to obtain analysis summaries of library code in the presence of callbacks.
Our algorithm preprocesses libraries in almost linear time, after which the contribution of the library in the complexity of the client analysis is (i)~linear in the number of call sites and (ii)~only logarithmic in the size of the whole library, as opposed to linear in the size of the whole library.
Finally, we prove that Dyck reachability is Boolean Matrix Multiplication-hard in general, and the hardness also holds for graphs of constant treewidth.
This hardness result strongly indicates that there exist no combinatorial algorithms for Dyck reachability with truly subcubic complexity.
Our third contribution is the formalization and algorithmic treatment of the Quantitative Interprocedural Analysis framework.
In this framework, the transitions of a recursive program are annotated as good, bad or neutral, and receive a weight which measures
the magnitude of their respective effect.
The Quantitative Interprocedural Analysis problem asks to determine whether there exists an infinite run of the program where the long-run ratio of the bad weights over the good weights is above a given threshold.
We illustrate how several quantitative problems related to static analysis of recursive programs can be instantiated in this framework,
and present some case studies to this direction.
Our fourth contribution is a new dynamic partial-order reduction for the stateless model checking of concurrent programs. Traditional approaches rely on the standard Mazurkiewicz equivalence between traces, by means of partitioning the trace space into equivalence classes, and attempting to explore a few representatives from each class.
We present a new dynamic partial-order reduction method called the Data-centric Partial Order Reduction (DC-DPOR).
Our algorithm is based on a new equivalence between traces, called the observation equivalence.
DC-DPOR explores a coarser partitioning of the trace space than any exploration method based on the standard Mazurkiewicz equivalence.
Depending on the program, the new partitioning can be even exponentially coarser.
Additionally, DC-DPOR spends only polynomial time in each explored class.
Our fifth contribution is the use of automata and game-theoretic verification techniques in the competitive analysis and synthesis of real-time scheduling algorithms for firm-deadline tasks.
On the analysis side, we leverage automata on infinite words to compute the competitive ratio of real-time schedulers subject to various environmental constraints.
On the synthesis side, we introduce a new instance of two-player mean-payoff partial-information games, and show
how the synthesis of an optimal real-time scheduler can be reduced to computing winning strategies in this new type of games
Goal-oriented adaptive multilevel stochastic Galerkin FEM
This paper is concerned with the numerical approximation of quantities of interest associated with solutions to parametric elliptic partial differential equations (PDEs). We consider a class of parametric elliptic PDEs where the underlying differential operator has affine dependence on a countably infinite number of uncertain parameters. We design a goal-oriented adaptive algorithm for approximating the functionals of solutions to this class of parametric PDEs. The algorithm applies to bounded linear goal functionals as well as to continuously Gateaux differentiable nonlinear functionals. In the algorithm, the approximations of parametric solutions to the primal and dual problems are computed using the multilevel stochastic Galerkin finite element method (SGFEM) and the adaptive refinement process is guided by reliable spatial and parametric error indicators that identify the dominant sources of error. We prove that the proposed algorithm generates multilevel SGFEM approximations for which the error estimates in the goal functional converge to zero. Furthermore, in the case of bounded linear goal functionals, we show that, under an appropriate saturation assumption, our goal-oriented adaptive strategy yields optimal convergence rates with respect to the overall dimension of the underlying multilevel approximations spaces
Recommended from our members
Optimization for Probabilistic Machine Learning
We have access to great variety of datasets more than any time in the history. Everyday, more data is collected from various natural resources and digital platforms. Great advances in the area of machine learning research in the past few decades have relied strongly on availability of these datasets. However, analyzing them imposes significant challenges that are mainly due to two factors. First, the datasets have complex structures with hidden interdependencies. Second, most of the valuable datasets are high dimensional and are largely scaled. The main goal of a machine learning framework is to design a model that is a valid representative of the observations and develop a learning algorithm to make inference about unobserved or latent data based on the observations. Discovering hidden patterns and inferring latent characteristics in such datasets is one of the greatest challenges in the area of machine learning research. In this dissertation, I will investigate some of the challenges in modeling and algorithm design, and present my research results on how to overcome these obstacles.
Analyzing data generally involves two main stages. The first stage is designing a model that is flexible enough to capture complex variation and latent structures in data and is robust enough to generalize well to the unseen data. Designing an expressive and interpretable model is one of crucial objectives in this stage. The second stage involves training learning algorithm on the observed data and measuring the accuracy of model and learning algorithm. This stage usually involves an optimization problem whose objective is to tune the model to the training data and learn the model parameters. Finding global optimal or sufficiently good local optimal solution is one of the main challenges in this step.
Probabilistic models are one of the best known models for capturing data generating process and quantifying uncertainties in data using random variables and probability distributions. They are powerful models that are shown to be adaptive and robust and can scale well to large datasets. However, most probabilistic models have a complex structure. Training them could become challenging commonly due to the presence of intractable integrals in the calculation. To remedy this, they require approximate inference strategies that often results in non-convex optimization problems. The optimization part ensures that the model is the best representative of data or data generating process. The non-convexity of an optimization problem take away the general guarantee on finding a global optimal solution. It will be shown later in this dissertation that inference for a significant number of probabilistic models require solving a non-convex optimization problem.
One of the well-known methods for approximate inference in probabilistic modeling is variational inference. In the Bayesian setting, the target is to learn the true posterior distribution for model parameters given the observations and prior distributions. The main challenge involves marginalization of all the other variables in the model except for the variable of interest. This high-dimensional integral is generally computationally hard, and for many models there is no known polynomial time algorithm for calculating them exactly. Variational inference deals with finding an approximate posterior distribution for Bayesian models where finding the true posterior distribution is analytically or numerically impossible. It assumes a family of distribution for the estimation, and finds the closest member of that family to the true posterior distribution using a distance measure. For many models though, this technique requires solving a non-convex optimization problem that has no general guarantee on reaching a global optimal solution. This dissertation presents a convex relaxation technique for dealing with hardness of the optimization involved in the inference.
The proposed convex relaxation technique is based on semidefinite optimization that has a general applicability to polynomial optimization problem. I will present theoretical foundations and in-depth details of this relaxation in this work. Linear dynamical systems represent the functionality of many real-world physical systems. They can describe the dynamics of a linear time-varying observation which is controlled by a controller unit with quadratic cost function objectives. Designing distributed and decentralized controllers is the goal of many of these systems, which computationally, results in a non-convex optimization problem. In this dissertation, I will further investigate the issues arising in this area and develop a convex relaxation framework to deal with the optimization challenges.
Setting the correct number of model parameters is an important aspect for a good probabilistic model. If there are only a few parameters, model may lack capturing all the essential relations and components in the observations while too many parameters may cause significant complications in learning or overfit to the observations. Non-parametric models are suitable techniques to deal with this issue. They allow the model to learn the appropriate number of parameters to describe the data and make predictions. In this dissertation, I will present my work on designing Bayesian non-parametric models as powerful tools for learning representations of data. Moreover, I will describe the algorithm that we derived to efficiently train the model on the observations and learn the number of model parameters.
Later in this dissertation, I will present my works on designing probabilistic models in combination with deep learning methods for representing sequential data. Sequential datasets comprise a significant portion of resources in the area of machine learning research. Designing models to capture dependencies in sequential datasets are of great interest and have a wide variety of applications in engineering, medicine and statistics. Recent advances in deep learning research has shown exceptional promises in this area. However, they lack interpretability in their general form. To remedy this, I will present my work on mixing probabilistic models with neural network models that results in better performance and expressiveness of the results