352 research outputs found
Pixel-wise Orthogonal Decomposition for Color Illumination Invariant and Shadow-free Image
In this paper, we propose a novel, effective and fast method to obtain a
color illumination invariant and shadow-free image from a single outdoor image.
Different from state-of-the-art methods for shadow-free image that either need
shadow detection or statistical learning, we set up a linear equation set for
each pixel value vector based on physically-based shadow invariants, deduce a
pixel-wise orthogonal decomposition for its solutions, and then get an
illumination invariant vector for each pixel value vector on an image. The
illumination invariant vector is the unique particular solution of the linear
equation set, which is orthogonal to its free solutions. With this illumination
invariant vector and Lab color space, we propose an algorithm to generate a
shadow-free image which well preserves the texture and color information of the
original image. A series of experiments on a diverse set of outdoor images and
the comparisons with the state-of-the-art methods validate our method.Comment: This paper has been published in Optics Express, Vol. 23, Issue 3,
pp. 2220-2239. The final version is available on
http://dx.doi.org/10.1364/OE.23.002220. Please refer to that version when
citing this pape
Efficient 3D object recognition via geometric information preservation
© 2019 Elsevier Ltd Accurate 3D object recognition and 6-DOF pose estimation have been pervasively applied to a variety of applications, such as unmanned warehouse, cooperative robots, and manufacturing industry. How to extract a robust and representative feature from the point clouds is an inevitable and important issue. In this paper, an unsupervised feature learning network is introduced to extract 3D keypoint features from point clouds directly, rather than transforming point clouds to voxel grids or projected RGB images, which saves computational time while preserving the object geometric information as well. Specifically, the proposed network features in a stacked point feature encoder, which can stack the local discriminative features within its neighborhoods to the original point-wise feature counterparts. The main framework consists of both offline training phase and online testing phase. In the offline training phase, the stacked point feature encoder is trained first and then generate feature database of all keypoints, which are sampled from synthetic point clouds of multiple model views. In the online testing phase, each feature extracted from the unknown testing scene is matched among the database by using the K-D tree voting strategy. Afterwards, the matching results are achieved by using the hypothesis & verification strategy. The proposed method is extensively evaluated on four public datasets and the results show that ours deliver comparable or even superior performances than the state-of-the-arts in terms of F1-score, Average of the 3D distance (ADD) and Recognition rate
CCR: Facial Image Editing with Continuity, Consistency and Reversibility
Three problems exist in sequential facial image editing: incontinuous
editing, inconsistent editing, and irreversible editing. Incontinuous editing
is that the current editing can not retain the previously edited attributes.
Inconsistent editing is that swapping the attribute editing orders can not
yield the same results. Irreversible editing means that operating on a facial
image is irreversible, especially in sequential facial image editing. In this
work, we put forward three concepts and corresponding definitions: editing
continuity, consistency, and reversibility. Then, we propose a novel model to
achieve the goal of editing continuity, consistency, and reversibility. A
sufficient criterion is defined to determine whether a model is continuous,
consistent, and reversible. Extensive qualitative and quantitative experimental
results validate our proposed model and show that a continuous, consistent and
reversible editing model has a more flexible editing function while preserving
facial identity. Furthermore, we think that our proposed definitions and model
will have wide and promising applications in multimedia processing. Code and
data are available at https://github.com/mickoluan/CCR.Comment: 10 pages, 11 figure
Low-Tubal-Rank Tensor Recovery via Factorized Gradient Descent
This paper considers the problem of recovering a tensor with an underlying
low-tubal-rank structure from a small number of corrupted linear measurements.
Traditional approaches tackling such a problem require the computation of
tensor Singular Value Decomposition (t-SVD), that is a computationally
intensive process, rendering them impractical for dealing with large-scale
tensors. Aim to address this challenge, we propose an efficient and effective
low-tubal-rank tensor recovery method based on a factorization procedure akin
to the Burer-Monteiro (BM) method. Precisely, our fundamental approach involves
decomposing a large tensor into two smaller factor tensors, followed by solving
the problem through factorized gradient descent (FGD). This strategy eliminates
the need for t-SVD computation, thereby reducing computational costs and
storage requirements. We provide rigorous theoretical analysis to ensure the
convergence of FGD under both noise-free and noisy situations. Additionally, it
is worth noting that our method does not require the precise estimation of the
tensor tubal-rank. Even in cases where the tubal-rank is slightly
overestimated, our approach continues to demonstrate robust performance. A
series of experiments have been carried out to demonstrate that, as compared to
other popular ones, our approach exhibits superior performance in multiple
scenarios, in terms of the faster computational speed and the smaller
convergence error.Comment: 13 pages, 4 figure
- …