253,979 research outputs found
Adaptive approximation of monotone functions
We study the classical problem of approximating a non-decreasing function in norm by sequentially querying its
values, for known compact real intervals , and a
known probability measure on \cX. For any function~ we characterize
the minimum number of evaluations of that algorithms need to guarantee an
approximation with an error below after
stopping. Unlike worst-case results that hold uniformly over all , our
complexity measure is dependent on each specific function . To address this
problem, we introduce GreedyBox, a generalization of an algorithm originally
proposed by Novak (1992) for numerical integration. We prove that GreedyBox
achieves an optimal sample complexity for any function , up to logarithmic
factors. Additionally, we uncover results regarding piecewise-smooth functions.
Perhaps as expected, the error of GreedyBox decreases much faster
for piecewise- functions than predicted by the algorithm (without any
knowledge on the smoothness of ). A simple modification even achieves
optimal minimax approximation rates for such functions, which we compute
explicitly. In particular, our findings highlight multiple performance gaps
between adaptive and non-adaptive algorithms, smooth and piecewise-smooth
functions, as well as monotone or non-monotone functions. Finally, we provide
numerical experiments to support our theoretical results
Finding Community Structure with Performance Guarantees in Complex Networks
Many networks including social networks, computer networks, and biological
networks are found to divide naturally into communities of densely connected
individuals. Finding community structure is one of fundamental problems in
network science. Since Newman's suggestion of using \emph{modularity} as a
measure to qualify the goodness of community structures, many efficient methods
to maximize modularity have been proposed but without a guarantee of
optimality. In this paper, we propose two polynomial-time algorithms to the
modularity maximization problem with theoretical performance guarantees. The
first algorithm comes with a \emph{priori guarantee} that the modularity of
found community structure is within a constant factor of the optimal modularity
when the network has the power-law degree distribution. Despite being mainly of
theoretical interest, to our best knowledge, this is the first approximation
algorithm for finding community structure in networks. In our second algorithm,
we propose a \emph{sparse metric}, a substantially faster linear programming
method for maximizing modularity and apply a rounding technique based on this
sparse metric with a \emph{posteriori approximation guarantee}. Our experiments
show that the rounding algorithm returns the optimal solutions in most cases
and are very scalable, that is, it can run on a network of a few thousand nodes
whereas the LP solution in the literature only ran on a network of at most 235
nodes
ベイジアンネットワークにおける確率推論の高速化のための最適三角化アルゴリズムの提案
Bayesian networks are widely used probabilistic graphical models that provide a compact representation of joint probability distributions over a set of variables. A common inference task in Bayesian networks is to compute the posterior marginal distributions for the unobserved variables given some evidence variables that we have already observed. However, the inference problem is known to be NP-hard and this complexity of inference limits the usage of Bayesian networks. Many attempts to improve the inference algorithm have been made in the past two decades. Currently, the junction tree algorithm is among the most prominent exact inference algorithms. To perform efficient inference on a Bayesian network using the junction tree algorithm, it is necessary to find a triangulation of the moral graph of the Bayesian network such that the total table size is small. In this context, the total table size is used to measure the computational complexity of the junction tree inference algorithm. This thesis focuses on exact algorithms for finding a triangulation that minimizes the total table size for a given Bayesian network. For optimal triangulation, Ottosen and Vomlel have proposed a depth-first search (DFS) algorithm. They also introduced several techniques to improve the DFS algorithm, including dynamic clique maintenance and coalescing map pruning. Nevertheless, the efficiency and scalability of their algorithm leave much room for improvement. First, the dynamic clique maintenance allows the recomputation of some cliques. Second, for a Bayesian network with n variables, the DFS algorithm runs in O*(n!) time because it explores a search space of all elimination orders. To mitigate these problems, an extended depth-first search (EDFS) algorithm is proposed in this thesis. The new EDFS algorithm introduces two techniques: (1) a new dynamic clique maintenance algorithm that computes only those cliques that contain a new edge, and (2) a new pruning rule, called pivot clique pruning. The new dynamic clique maintenance algorithm explores a smaller search space and runs faster than the Ottosen and Vomlel approach. This improvement can decrease the overhead cost of the DFS algorithm, and the pivot clique pruning reduces the size of the search space by a factor of O(n2). Our empirical results show that our proposed algorithm finds an optimal triangulation markedly faster than the state-of-the-art algorithm does.電気通信大学201
Using Multi Population Cultural Algorithms to prune Deep Neural Networks
The success of Deep Neural Networks (DNN) in classification is accompanied by a drastic increase in weight parameters which also increases the computational and storage costs. Pruning of DNN involves identifying and removing redundant parameters with little or no loss of accuracy. Layer-wise pruning of weights by their magnitude has shown to be an efficient method to prune neural networks. However, finding the optimal values of the threshold for each layer is a challenging task given the large search space. To solve this problem, we use multi population cultural algorithm which is an evolutionary algorithm that takes advantage of knowledge domains and faster convergence and is used in many optimization problems. We experiment it on LeNet-style models and measure the level of pruning through the pruning ratio. Results show that our method achieves the best pruning ratio (864 on LeNet5) compared with some state-of-the-art DNN pruning methods. By removing redundant parameters, the computational and storage costs are reduced significantly
The Fragility of Optimized Bandit Algorithms
Much of the literature on optimal design of bandit algorithms is based on
minimization of expected regret. It is well known that designs that are optimal
over certain exponential families can achieve expected regret that grows
logarithmically in the number of arm plays, at a rate governed by the
Lai-Robbins lower bound. In this paper, we show that when one uses such
optimized designs, the regret distribution of the associated algorithms
necessarily has a very heavy tail, specifically, that of a truncated Cauchy
distribution. Furthermore, for , the 'th moment of the regret
distribution grows much faster than poly-logarithmically, in particular as a
power of the total number of arm plays. We show that optimized UCB bandit
designs are also fragile in an additional sense, namely when the problem is
even slightly mis-specified, the regret can grow much faster than the
conventional theory suggests. Our arguments are based on standard
change-of-measure ideas, and indicate that the most likely way that regret
becomes larger than expected is when the optimal arm returns below-average
rewards in the first few arm plays, thereby causing the algorithm to believe
that the arm is sub-optimal. To alleviate the fragility issues exposed, we show
that UCB algorithms can be modified so as to ensure a desired degree of
robustness to mis-specification. In doing so, we also provide a sharp trade-off
between the amount of UCB exploration and the tail exponent of the resulting
regret distribution
Static mapping heuristics for tasks with dependencies, priorities, deadlines, and multiple versions in heterogeneous environments
Includes bibliographical references.Heterogeneous computing (HC) environments composed of interconnected machines with varied computational capabilities are well suited to meet the computational demands of large, diverse groups of tasks. The problem of mapping (defined as matching and scheduling) these tasks onto the machines of a distributed HC environment has been shown, in general, to be NP-complete. Therefore, the development of heuristic techniques to find near-optimal solutions is required. In the HC environment investigated, tasks had deadlines, priorities, multiple versions, and may be composed of communicating subtasks. The best static (off-line) techniques from some previous studies were adapted and applied to this mapping problem: a genetic algorithm (GA), a GENITOR-style algorithm, and a greedy Min-min technique. Simulation studies compared the performance of these heuristics in several overloaded scenarios, i.e., not all tasks executed. The performance measure used was a sum of weighted priorities of tasks that completed before their deadline, adjusted based on the version of the task used. It is shown that for the cases studied here, the GENITOR technique found the best results, but the faster Min-min approach also performed very well.This research was supported in part by the DARPA/ITO Quorum Program under GSA subcontract number GS09K99BH0250 and a Purdue University Dean of Engineering Donnan Scholarship
- …