2,635 research outputs found
Doctor of Philosophy
dissertationImage segmentation entails the partitioning of an image domain, usually two or three dimensions, so that each partition or segment has some meaning that is relevant to the application at hand. Accurate image segmentation is a crucial challenge in many disciplines, including medicine, computer vision, and geology. In some applications, heterogeneous pixel intensities; noisy, ill-defined, or diffusive boundaries; and irregular shapes with high variability can make it challenging to meet accuracy requirements. Various segmentation approaches tackle such challenges by casting the segmentation problem as an energy-minimization problem, and solving it using efficient optimization algorithms. These approaches are broadly classified as either region-based or edge (surface)-based depending on the features on which they operate. The focus of this dissertation is on the development of a surface-based energy model, the design of efficient formulations of optimization frameworks to incorporate such energy, and the solution of the energy-minimization problem using graph cuts. This dissertation utilizes a set of four papers whose motivation is the efficient extraction of the left atrium wall from the late gadolinium enhancement magnetic resonance imaging (LGE-MRI) image volume. This dissertation utilizes these energy formulations for other applications, including contact lens segmentation in the optical coherence tomography (OCT) data and the extraction of geologic features in seismic data. Chapters 2 through 5 (papers 1 through 4) explore building a surface-based image segmentation model by progressively adding components to improve its accuracy and robustness. The first paper defines a parametric search space and its discrete formulation in the form of a multilayer three-dimensional mesh model within which the segmentation takes place. It includes a generative intensity model, and we optimize using a graph formulation of the surface net problem. The second paper proposes a Bayesian framework with a Markov random field (MRF) prior that gives rise to another class of surface nets, which provides better segmentation with smooth boundaries. The third paper presents a maximum a posteriori (MAP)-based surface estimation framework that relies on a generative image model by incorporating global shape priors, in addition to the MRF, within the Bayesian formulation. Thus, the resulting surface not only depends on the learned model of shapes,but also accommodates the test data irregularities through smooth deviations from these priors. Further, the paper proposes a new shape parameter estimation scheme, in closed form, for segmentation as a part of the optimization process. Finally, the fourth paper (under review at the time of this document) presents an extensive analysis of the MAP framework and presents improved mesh generation and generative intensity models. It also performs a thorough analysis of the segmentation results that demonstrates the effectiveness of the proposed method qualitatively, quantitatively, and clinically. Chapter 6, consisting of unpublished work, demonstrates the application of an MRF-based Bayesian framework to segment coupled surfaces of contact lenses in optical coherence tomography images. This chapter also shows an application related to the extraction of geological structures in seismic volumes. Due to the large sizes of seismic volume datasets, we also present fast, approximate surface-based energy minimization strategies that achieve better speed-ups and memory consumption
A Neighborhood-preserving Graph Summarization
We introduce in this paper a new summarization method for large graphs. Our
summarization approach retains only a user-specified proportion of the
neighbors of each node in the graph. Our main aim is to simplify large graphs
so that they can be analyzed and processed effectively while preserving as many
of the node neighborhood properties as possible. Since many graph algorithms
are based on the neighborhood information available for each node, the idea is
to produce a smaller graph which can be used to allow these algorithms to
handle large graphs and run faster while providing good approximations.
Moreover, our compression allows users to control the size of the compressed
graph by adjusting the amount of information loss that can be tolerated. The
experiments conducted on various real and synthetic graphs show that our
compression reduces considerably the size of the graphs. Moreover, we conducted
several experiments on the obtained summaries using various graph algorithms
and applications, such as node embedding, graph classification and shortest
path approximations. The obtained results show interesting trade-offs between
the algorithms runtime speed-up and the precision loss.Comment: 17 pages, 10 figure
Regenerator Location Problem and survivable extensions: A hub covering location perspective
Cataloged from PDF version of article.In a telecommunications network the reach of an optical signal is the maximum distance it can traverse before its quality degrades. Regenerators are devices to extend the optical reach. The regenerator placement problem seeks to place the minimum number of regenerators in an optical network so as to facilitate the communication of a signal between any node pair. In this study, the Regenerator Location Problem is revisited from the hub location perspective directing our focus to applications arising in transportation settings. Two new dimensions involving the challenges of survivability are introduced to the problem. Under partial survivability, our designs hedge against failures in the regeneration equipment only, whereas under full survivability failures on any of the network nodes are accounted for by the utilization of extra regeneration equipment. All three variations of the problem are studied in a unifying framework involving the introduction of individual flow-based compact formulations as well as cut formulations and the implementation of branch and cut algorithms based on the cut formulations. Extensive computational experiments are conducted in order to evaluate the performance of the proposed solution methodologies and to gain insights from realistic instances. (C) 2014 Elsevier Ltd. All rights reserved
Convex Graph Invariant Relaxations For Graph Edit Distance
The edit distance between two graphs is a widely used measure of similarity
that evaluates the smallest number of vertex and edge deletions/insertions
required to transform one graph to another. It is NP-hard to compute in
general, and a large number of heuristics have been proposed for approximating
this quantity. With few exceptions, these methods generally provide upper
bounds on the edit distance between two graphs. In this paper, we propose a new
family of computationally tractable convex relaxations for obtaining lower
bounds on graph edit distance. These relaxations can be tailored to the
structural properties of the particular graphs via convex graph invariants.
Specific examples that we highlight in this paper include constraints on the
graph spectrum as well as (tractable approximations of) the stability number
and the maximum-cut values of graphs. We prove under suitable conditions that
our relaxations are tight (i.e., exactly compute the graph edit distance) when
one of the graphs consists of few eigenvalues. We also validate the utility of
our framework on synthetic problems as well as real applications involving
molecular structure comparison problems in chemistry.Comment: 27 pages, 7 figure
- …