26,119 research outputs found
A Message-Passing Algorithm for Counting Short Cycles in a Graph
A message-passing algorithm for counting short cycles in a graph is
presented. For bipartite graphs, which are of particular interest in coding,
the algorithm is capable of counting cycles of length g, g +2,..., 2g - 2,
where g is the girth of the graph. For a general (non-bipartite) graph, cycles
of length g; g + 1, ..., 2g - 1 can be counted. The algorithm is based on
performing integer additions and subtractions in the nodes of the graph and
passing extrinsic messages to adjacent nodes. The complexity of the proposed
algorithm grows as , where is the number of edges in the
graph. For sparse graphs, the proposed algorithm significantly outperforms the
existing algorithms in terms of computational complexity and memory
requirements.Comment: Submitted to IEEE Trans. Inform. Theory, April 21, 2010
Neighbourhood-consensus message passing and its potentials in image processing applications
In this paper, a novel algorithm for inference in Markov Random Fields (MRFs) is presented. Its goal is to find approximate maximum a posteriori estimates in a simple manner by combining neighbourhood influence of iterated conditional modes (ICM) and message passing of loopy belief propagation (LBP). We call the proposed method neighbourhood-consensus message passing because a single joint message is sent from the specified neighbourhood to the central node. The message, as a function of beliefs, represents the agreement of all nodes within the neighbourhood regarding the labels of the central node. This way we are able to overcome the disadvantages of reference algorithms, ICM and LBP. On one hand, more information is propagated in comparison with ICM, while on the other hand, the huge amount of pairwise interactions is avoided in comparison with LBP by working with neighbourhoods. The idea is related to the previously developed iterated conditional expectations algorithm. Here we revisit it and redefine it in a message passing framework in a more general form. The results on three different benchmarks demonstrate that the proposed technique can perform well both for binary and multi-label MRFs without any limitations on the model definition. Furthermore, it manifests improved performance over related techniques either in terms of quality and/or speed
Binary Linear Classification and Feature Selection via Generalized Approximate Message Passing
For the problem of binary linear classification and feature selection, we
propose algorithmic approaches to classifier design based on the generalized
approximate message passing (GAMP) algorithm, recently proposed in the context
of compressive sensing. We are particularly motivated by problems where the
number of features greatly exceeds the number of training examples, but where
only a few features suffice for accurate classification. We show that
sum-product GAMP can be used to (approximately) minimize the classification
error rate and max-sum GAMP can be used to minimize a wide variety of
regularized loss functions. Furthermore, we describe an
expectation-maximization (EM)-based scheme to learn the associated model
parameters online, as an alternative to cross-validation, and we show that
GAMP's state-evolution framework can be used to accurately predict the
misclassification rate. Finally, we present a detailed numerical study to
confirm the accuracy, speed, and flexibility afforded by our GAMP-based
approaches to binary linear classification and feature selection
- …