160 research outputs found
Who clicks there!: Anonymizing the photographer in a camera saturated society
In recent years, social media has played an increasingly important role in
reporting world events. The publication of crowd-sourced photographs and videos
in near real-time is one of the reasons behind the high impact. However, the
use of a camera can draw the photographer into a situation of conflict.
Examples include the use of cameras by regulators collecting evidence of Mafia
operations; citizens collecting evidence of corruption at a public service
outlet; and political dissidents protesting at public rallies. In all these
cases, the published images contain fairly unambiguous clues about the location
of the photographer (scene viewpoint information). In the presence of adversary
operated cameras, it can be easy to identify the photographer by also combining
leaked information from the photographs themselves. We call this the camera
location detection attack. We propose and review defense techniques against
such attacks. Defenses such as image obfuscation techniques do not protect
camera-location information; current anonymous publication technologies do not
help either. However, the use of view synthesis algorithms could be a promising
step in the direction of providing probabilistic privacy guarantees
Ensemble of Example-Dependent Cost-Sensitive Decision Trees
Several real-world classification problems are example-dependent
cost-sensitive in nature, where the costs due to misclassification vary between
examples and not only within classes. However, standard classification methods
do not take these costs into account, and assume a constant cost of
misclassification errors. In previous works, some methods that take into
account the financial costs into the training of different algorithms have been
proposed, with the example-dependent cost-sensitive decision tree algorithm
being the one that gives the highest savings. In this paper we propose a new
framework of ensembles of example-dependent cost-sensitive decision-trees. The
framework consists in creating different example-dependent cost-sensitive
decision trees on random subsamples of the training set, and then combining
them using three different combination approaches. Moreover, we propose two new
cost-sensitive combination approaches; cost-sensitive weighted voting and
cost-sensitive stacking, the latter being based on the cost-sensitive logistic
regression method. Finally, using five different databases, from four
real-world applications: credit card fraud detection, churn modeling, credit
scoring and direct marketing, we evaluate the proposed method against
state-of-the-art example-dependent cost-sensitive techniques, namely,
cost-proportionate sampling, Bayes minimum risk and cost-sensitive decision
trees. The results show that the proposed algorithms have better results for
all databases, in the sense of higher savings.Comment: 13 pages, 6 figures, Submitted for possible publicatio
PvDeConv: Point-Voxel Deconvolution for Autoencoding CAD Construction in 3D
We propose a Point-Voxel DeConvolution (PVDeConv) module for 3D data
autoencoder. To demonstrate its efficiency we learn to synthesize
high-resolution point clouds of 10k points that densely describe the underlying
geometry of Computer Aided Design (CAD) models. Scanning artifacts, such as
protrusions, missing parts, smoothed edges and holes, inevitably appear in real
3D scans of fabricated CAD objects. Learning the original CAD model
construction from a 3D scan requires a ground truth to be available together
with the corresponding 3D scan of an object. To solve the gap, we introduce a
new dedicated dataset, the CC3D, containing 50k+ pairs of CAD models and their
corresponding 3D meshes. This dataset is used to learn a convolutional
autoencoder for point clouds sampled from the pairs of 3D scans - CAD models.
The challenges of this new dataset are demonstrated in comparison with other
generative point cloud sampling models trained on ShapeNet. The CC3D
autoencoder is efficient with respect to memory consumption and training time
as compared to stateof-the-art models for 3D data generation.Comment: 2020 IEEE International Conference on Image Processing (ICIP
Fast Adaptive Reparametrization (FAR) with Application to Human Action Recognition
In this paper, a fast approach for curve reparametrization, called Fast Adaptive Reparamterization (FAR), is introduced. Instead of computing an optimal matching between two curves such as Dynamic Time Warping (DTW) and elastic distance-based approaches, our method is applied to each curve independently, leading to linear computational complexity. It is based on a simple replacement of the curve parameter by a variable invariant under specific variations of reparametrization. The choice of this variable is heuristically made according to the application of interest. In addition to being fast, the proposed reparametrization can be applied not only to curves observed in Euclidean spaces but also to feature curves living in Riemannian spaces. To validate our approach, we apply it to the scenario of human action recognition using curves living in the Riemannian product Special Euclidean space SE(3) n. The obtained results on three benchmarks for human action recognition (MSRAction3D, Florence3D, and UTKinect) show that our approach competes with state-of-the-art methods in terms of accuracy and computational cost
Towards Generalization of 3D Human Pose Estimation In The Wild
In this paper, we propose 3DBodyTex.Pose, a dataset that addresses the task
of 3D human pose estimation in-the-wild. Generalization to in-the-wild images
remains limited due to the lack of adequate datasets. Existent ones are usually
collected in indoor controlled environments where motion capture systems are
used to obtain the 3D ground-truth annotations of humans. 3DBodyTex.Pose offers
high quality and rich data containing 405 different real subjects in various
clothing and poses, and 81k image samples with ground-truth 2D and 3D pose
annotations. These images are generated from 200 viewpoints among which 70
challenging extreme viewpoints. This data was created starting from high
resolution textured 3D body scans and by incorporating various realistic
backgrounds. Retraining a state-of-the-art 3D pose estimation approach using
data augmented with 3DBodyTex.Pose showed promising improvement in the overall
performance, and a sensible decrease in the per joint position error when
testing on challenging viewpoints. The 3DBodyTex.Pose is expected to offer the
research community with new possibilities for generalizing 3D pose estimation
from monocular in-the-wild images
DELO: Deep Evidential LiDAR Odometry using Partial Optimal Transport
Accurate, robust, and real-time LiDAR-based odometry (LO) is imperative for
many applications like robot navigation, globally consistent 3D scene map
reconstruction, or safe motion-planning. Though LiDAR sensor is known for its
precise range measurement, the non-uniform and uncertain point sampling density
induce structural inconsistencies. Hence, existing supervised and unsupervised
point set registration methods fail to establish one-to-one matching
correspondences between LiDAR frames. We introduce a novel deep learning-based
real-time (approx. 35-40ms per frame) LO method that jointly learns accurate
frame-to-frame correspondences and model's predictive uncertainty (PU) as
evidence to safe-guard LO predictions. In this work, we propose (i) partial
optimal transportation of LiDAR feature descriptor for robust LO estimation,
(ii) joint learning of predictive uncertainty while learning odometry over
driving sequences, and (iii) demonstrate how PU can serve as evidence for
necessary pose-graph optimization when LO network is either under or over
confident. We evaluate our method on KITTI dataset and show competitive
performance, even superior generalization ability over recent state-of-the-art
approaches. Source codes are available.Comment: Accepted in ICCV 2023 Worksho
Deformation Based 3D Facial Expression Representation
We propose a deformation based representation for analyzing expressions from 3D faces. A point cloud of a 3D face is decomposed into an ordered deformable set of curves that start from a fixed point. Subsequently, a mapping function is defined to identify the set of curves with an element of a high dimensional matrix Lie group, specifically the direct product of SE(3). Representing 3D faces as an element of a high dimensional Lie group has two main advantages. First, using the group structure, facial expressions can be decoupled from a neutral face. Second, an underlying non-linear facial expression manifold can be captured with the Lie group and mapped to a linear space, Lie algebra of the group. This opens up the possibility of classifying facial expressions with linear models without compromising the underlying manifold. Alternatively, linear combinations of linearised facial expressions can be mapped back from the Lie algebra to the Lie group. The approach is tested on the BU-3DFE and the Bosphorus datasets. The results show that the proposed approach performed comparably, on the BU-3DFE dataset, without using features or extensive landmark points
- …