109,109 research outputs found
PENERAPAN ALGORITMA DIJKSTRA DALAM PENENTUAN LINTASAN TERPENDEK MENUJU UPT. PUSKESMAS CILODONG KOTA DEPOK
One of the government's efforts in providing health to the community is the construction of health centers in each sub-district, and the community is expected to be able to take advantage of the health facilities provided by the government. One of the problems that exist in the community is determining the shortest distance to the puskesmas. In Depok City, there are 26 routes that can be passed from the 38 nodes or vertices to the Cilodong Health Center with the starting point of the Depok mayor's office. This study uses a survey research method to calculate the actual distance at each node or vertex, the purpose of this study is to determine the shortest path taken by the starting point from the Depok mayor's office to get to the Cilodong Health Center by applying the dijkstra algorithm. This dijkstra algorithm works by visiting all existing points and making a route if there are 2 routes to the same 1 point then the route that has the lowest weight is chosen so that all points have an optimal route. This quest continues until the final destination point. After doing this research and testing using a simple application to calculate the distance by applying the djikstra algorithm, it was found that the shortest path taken to the destination is through the GDC Main Gate or on the test results in Iteration 26. From the results of this study, people can choose this closest route to save time when viewed from the distance of the existing track. For further research, it is expected to be able to compare two other algorithms and other parameters so that the closest route with the fastest time is obtained
Improvements on the k-center problem for uncertain data
In real applications, there are situations where we need to model some
problems based on uncertain data. This leads us to define an uncertain model
for some classical geometric optimization problems and propose algorithms to
solve them. In this paper, we study the -center problem, for uncertain
input. In our setting, each uncertain point is located independently from
other points in one of several possible locations in a metric space with metric , with specified probabilities
and the goal is to compute -centers that minimize the
following expected cost here
is the probability space of all realizations of given uncertain points and
In restricted assigned version of this problem, an assignment is given for any choice of centers and the
goal is to minimize In unrestricted version, the
assignment is not specified and the goal is to compute centers
and an assignment that minimize the above expected
cost.
We give several improved constant approximation factor algorithms for the
assigned versions of this problem in a Euclidean space and in a general metric
space. Our results significantly improve the results of \cite{guh} and
generalize the results of \cite{wang} to any dimension. Our approach is to
replace a certain center point for each uncertain point and study the
properties of these certain points. The proposed algorithms are efficient and
simple to implement
Net and Prune: A Linear Time Algorithm for Euclidean Distance Problems
We provide a general framework for getting expected linear time constant
factor approximations (and in many cases FPTAS's) to several well known
problems in Computational Geometry, such as -center clustering and farthest
nearest neighbor. The new approach is robust to variations in the input
problem, and yet it is simple, elegant and practical. In particular, many of
these well studied problems which fit easily into our framework, either
previously had no linear time approximation algorithm, or required rather
involved algorithms and analysis. A short list of the problems we consider
include farthest nearest neighbor, -center clustering, smallest disk
enclosing points, th largest distance, th smallest -nearest
neighbor distance, th heaviest edge in the MST and other spanning forest
type problems, problems involving upward closed set systems, and more. Finally,
we show how to extend our framework such that the linear running time bound
holds with high probability
Minimizing the average distance to a closest leaf in a phylogenetic tree
When performing an analysis on a collection of molecular sequences, it can be
convenient to reduce the number of sequences under consideration while
maintaining some characteristic of a larger collection of sequences. For
example, one may wish to select a subset of high-quality sequences that
represent the diversity of a larger collection of sequences. One may also wish
to specialize a large database of characterized "reference sequences" to a
smaller subset that is as close as possible on average to a collection of
"query sequences" of interest. Such a representative subset can be useful
whenever one wishes to find a set of reference sequences that is appropriate to
use for comparative analysis of environmentally-derived sequences, such as for
selecting "reference tree" sequences for phylogenetic placement of metagenomic
reads. In this paper we formalize these problems in terms of the minimization
of the Average Distance to the Closest Leaf (ADCL) and investigate algorithms
to perform the relevant minimization. We show that the greedy algorithm is not
effective, show that a variant of the Partitioning Among Medoids (PAM)
heuristic gets stuck in local minima, and develop an exact dynamic programming
approach. Using this exact program we note that the performance of PAM appears
to be good for simulated trees, and is faster than the exact algorithm for
small trees. On the other hand, the exact program gives solutions for all
numbers of leaves less than or equal to the given desired number of leaves,
while PAM only gives a solution for the pre-specified number of leaves. Via
application to real data, we show that the ADCL criterion chooses chimeric
sequences less often than random subsets, while the maximization of
phylogenetic diversity chooses them more often than random. These algorithms
have been implemented in publicly available software.Comment: Please contact us with any comments or questions
Approximating -Median via Pseudo-Approximation
We present a novel approximation algorithm for -median that achieves an
approximation guarantee of
, improving upon the decade-old ratio of .
Our approach is based on two components, each of which, we believe, is of
independent interest.
First, we show that in order to give an -approximation algorithm for
-median, it is sufficient to give a \emph{pseudo-approximation algorithm}
that finds an -approximate solution by opening facilities.
This is a rather surprising result as there exist instances for which opening
facilities may lead to a significant smaller cost than if only
facilities were opened.
Second, we give such a pseudo-approximation algorithm with . Prior to our work, it was not even known whether opening
facilities would help improve the approximation ratio.Comment: 18 page
Fast Hierarchical Clustering and Other Applications of Dynamic Closest Pairs
We develop data structures for dynamic closest pair problems with arbitrary
distance functions, that do not necessarily come from any geometric structure
on the objects. Based on a technique previously used by the author for
Euclidean closest pairs, we show how to insert and delete objects from an
n-object set, maintaining the closest pair, in O(n log^2 n) time per update and
O(n) space. With quadratic space, we can instead use a quadtree-like structure
to achieve an optimal time bound, O(n) per update. We apply these data
structures to hierarchical clustering, greedy matching, and TSP heuristics, and
discuss other potential applications in machine learning, Groebner bases, and
local improvement algorithms for partition and placement problems. Experiments
show our new methods to be faster in practice than previously used heuristics.Comment: 20 pages, 9 figures. A preliminary version of this paper appeared at
the 9th ACM-SIAM Symp. on Discrete Algorithms, San Francisco, 1998, pp.
619-628. For source code and experimental results, see
http://www.ics.uci.edu/~eppstein/projects/pairs
Statistical Pruning for Near-Maximum Likelihood Decoding
In many communications problems, maximum-likelihood (ML) decoding reduces to finding the closest (skewed) lattice point in N-dimensions to a given point xisin CN. In its full generality, this problem is known to be NP-complete. Recently, the expected complexity of the sphere decoder, a particular algorithm that solves the ML problem exactly, has been computed. An asymptotic analysis of this complexity has also been done where it is shown that the required computations grow exponentially in N for any fixed SNR. At the same time, numerical computations of the expected complexity show that there are certain ranges of rates, SNRs and dimensions N for which the expected computation (counted as the number of scalar multiplications) involves no more than N3 computations. However, when the dimension of the problem grows too large, the required computations become prohibitively large, as expected from the asymptotic exponential complexity. In this paper, we propose an algorithm that, for large N, offers substantial computational savings over the sphere decoder, while maintaining performance arbitrarily close to ML. We statistically prune the search space to a subset that, with high probability, contains the optimal solution, thereby reducing the complexity of the search. Bounds on the error performance of the new method are proposed. The complexity of the new algorithm is analyzed through an upper bound. The asymptotic behavior of the upper bound for large N is also analyzed which shows that the upper bound is also exponential but much lower than the sphere decoder. Simulation results show that the algorithm is much more efficient than the original sphere decoder for smaller dimensions as well, and does not sacrifice much in terms of performance
- …