11,617 research outputs found
Confidence driven TGV fusion
We introduce a novel model for spatially varying variational data fusion,
driven by point-wise confidence values. The proposed model allows for the joint
estimation of the data and the confidence values based on the spatial coherence
of the data. We discuss the main properties of the introduced model as well as
suitable algorithms for estimating the solution of the corresponding biconvex
minimization problem and their convergence. The performance of the proposed
model is evaluated considering the problem of depth image fusion by using both
synthetic and real data from publicly available datasets
Collaborative Representation based Classification for Face Recognition
By coding a query sample as a sparse linear combination of all training
samples and then classifying it by evaluating which class leads to the minimal
coding residual, sparse representation based classification (SRC) leads to
interesting results for robust face recognition. It is widely believed that the
l1- norm sparsity constraint on coding coefficients plays a key role in the
success of SRC, while its use of all training samples to collaboratively
represent the query sample is rather ignored. In this paper we discuss how SRC
works, and show that the collaborative representation mechanism used in SRC is
much more crucial to its success of face classification. The SRC is a special
case of collaborative representation based classification (CRC), which has
various instantiations by applying different norms to the coding residual and
coding coefficient. More specifically, the l1 or l2 norm characterization of
coding residual is related to the robustness of CRC to outlier facial pixels,
while the l1 or l2 norm characterization of coding coefficient is related to
the degree of discrimination of facial features. Extensive experiments were
conducted to verify the face recognition accuracy and efficiency of CRC with
different instantiations.Comment: It is a substantial revision of a previous conference paper (L.
Zhang, M. Yang, et al. "Sparse Representation or Collaborative
Representation: Which Helps Face Recognition?" in ICCV 2011
Cross-Scale Cost Aggregation for Stereo Matching
Human beings process stereoscopic correspondence across multiple scales.
However, this bio-inspiration is ignored by state-of-the-art cost aggregation
methods for dense stereo correspondence. In this paper, a generic cross-scale
cost aggregation framework is proposed to allow multi-scale interaction in cost
aggregation. We firstly reformulate cost aggregation from a unified
optimization perspective and show that different cost aggregation methods
essentially differ in the choices of similarity kernels. Then, an inter-scale
regularizer is introduced into optimization and solving this new optimization
problem leads to the proposed framework. Since the regularization term is
independent of the similarity kernel, various cost aggregation methods can be
integrated into the proposed general framework. We show that the cross-scale
framework is important as it effectively and efficiently expands
state-of-the-art cost aggregation methods and leads to significant
improvements, when evaluated on Middlebury, KITTI and New Tsukuba datasets.Comment: To Appear in 2013 IEEE Conference on Computer Vision and Pattern
Recognition (CVPR). 2014 (poster, 29.88%
Nearfield Acoustic Holography using sparsity and compressive sampling principles
Regularization of the inverse problem is a complex issue when using
Near-field Acoustic Holography (NAH) techniques to identify the vibrating
sources. This paper shows that, for convex homogeneous plates with arbitrary
boundary conditions, new regularization schemes can be developed, based on the
sparsity of the normal velocity of the plate in a well-designed basis, i.e. the
possibility to approximate it as a weighted sum of few elementary basis
functions. In particular, these new techniques can handle discontinuities of
the velocity field at the boundaries, which can be problematic with standard
techniques. This comes at the cost of a higher computational complexity to
solve the associated optimization problem, though it remains easily tractable
with out-of-the-box software. Furthermore, this sparsity framework allows us to
take advantage of the concept of Compressive Sampling: under some conditions on
the sampling process (here, the design of a random array, which can be
numerically and experimentally validated), it is possible to reconstruct the
sparse signals with significantly less measurements (i.e., microphones) than
classically required. After introducing the different concepts, this paper
presents numerical and experimental results of NAH with two plate geometries,
and compares the advantages and limitations of these sparsity-based techniques
over standard Tikhonov regularization.Comment: Journal of the Acoustical Society of America (2012
- …