5,751 research outputs found
Object-based 2D-to-3D video conversion for effective stereoscopic content generation in 3D-TV applications
Three-dimensional television (3D-TV) has gained increasing popularity in the broadcasting domain, as it enables enhanced viewing experiences in comparison to conventional two-dimensional (2D) TV. However, its application has been constrained due to the lack of essential contents, i.e., stereoscopic videos. To alleviate such content shortage, an economical and practical solution is to reuse the huge media resources that are available in monoscopic 2D and convert them to stereoscopic 3D. Although stereoscopic video can be generated from monoscopic sequences using depth measurements extracted from cues like focus blur, motion and size, the quality of the resulting video may be poor as such measurements are usually arbitrarily defined and appear inconsistent with the real scenes. To help solve this problem, a novel method for object-based stereoscopic video generation is proposed which features i) optical-flow based occlusion reasoning in determining depth ordinal, ii) object segmentation using improved region-growing from masks of determined depth layers, and iii) a hybrid depth estimation scheme using content-based matching (inside a small library of true stereo image pairs) and depth-ordinal based regularization. Comprehensive experiments have validated the effectiveness of our proposed 2D-to-3D conversion method in generating stereoscopic videos of consistent depth measurements for 3D-TV applications
Component-wise modeling of articulated objects
We introduce a novel framework for modeling articulated objects based on the aspects of their components. By decomposing the object into components, we divide the problem in smaller modeling tasks. After obtaining 3D models for each component aspect by employing a shape deformation paradigm, we merge them together, forming the object components. The final model is obtained by assembling the components using an optimization scheme which fits the respective 3D models to the corresponding apparent contours in a reference pose. The results suggest that our approach can produce realistic 3D models of articulated objects in reasonable time
Reproducibility of the dynamics of facial expressions in unilateral facial palsy
The aim of this study was to assess the reproducibility of non-verbal facial
expressions in unilateral facial paralysis using dynamic four-dimensional (4D)
imaging. The Di4D system was used to record five facial expressions of 20 adult
patients. The system captured 60 three-dimensional (3D) images per second; each
facial expression took 3–4 seconds which was recorded in real time. Thus a set of
180 3D facial images was generated for each expression. The procedure was
repeated after 30 min to assess the reproducibility of the expressions. A
mathematical facial mesh consisting of thousands of quasi-point ‘vertices’ was
conformed to the face in order to determine the morphological characteristics in a
comprehensive manner. The vertices were tracked throughout the sequence of the
180 images. Five key 3D facial frames from each sequence of images were
analyzed. Comparisons were made between the first and second capture of each
facial expression to assess the reproducibility of facial movements. Corresponding
images were aligned using partial Procrustes analysis, and the root mean square
distance between them was calculated and analyzed statistically (paired Student ttest,
P < 0.05). Facial expressions of lip purse, cheek puff, and raising of eyebrows
were reproducible. Facial expressions of maximum smile and forceful eye closure
were not reproducible. The limited coordination of various groups of facial muscles
contributed to the lack of reproducibility of these facial expressions. 4D imaging is a
useful clinical tool for the assessment of facial expressions
CUR Decompositions, Similarity Matrices, and Subspace Clustering
A general framework for solving the subspace clustering problem using the CUR
decomposition is presented. The CUR decomposition provides a natural way to
construct similarity matrices for data that come from a union of unknown
subspaces . The similarity
matrices thus constructed give the exact clustering in the noise-free case.
Additionally, this decomposition gives rise to many distinct similarity
matrices from a given set of data, which allow enough flexibility to perform
accurate clustering of noisy data. We also show that two known methods for
subspace clustering can be derived from the CUR decomposition. An algorithm
based on the theoretical construction of similarity matrices is presented, and
experiments on synthetic and real data are presented to test the method.
Additionally, an adaptation of our CUR based similarity matrices is utilized
to provide a heuristic algorithm for subspace clustering; this algorithm yields
the best overall performance to date for clustering the Hopkins155 motion
segmentation dataset.Comment: Approximately 30 pages. Current version contains improved algorithm
and numerical experiments from the previous versio
Disparity map generation based on trapezoidal camera architecture for multiview video
Visual content acquisition is a strategic functional block of any visual system. Despite its wide possibilities,
the arrangement of cameras for the acquisition of good quality visual content for use in multi-view video
remains a huge challenge. This paper presents the mathematical description of trapezoidal camera
architecture and relationships which facilitate the determination of camera position for visual content
acquisition in multi-view video, and depth map generation. The strong point of Trapezoidal Camera
Architecture is that it allows for adaptive camera topology by which points within the scene, especially the
occluded ones can be optically and geometrically viewed from several different viewpoints either on the
edge of the trapezoid or inside it. The concept of maximum independent set, trapezoid characteristics, and
the fact that the positions of cameras (with the exception of few) differ in their vertical coordinate
description could very well be used to address the issue of occlusion which continues to be a major
problem in computer vision with regards to the generation of depth map
Flight of the dragonflies and damselflies
This work is a synthesis of our current understanding of the mechanics, aerodynamics and visually mediated control of dragonfly and damselfly flight, with the addition of new experimental and computational data in several key areas. These are: the diversity of dragonfly wing morphologies, the aerodynamics of gliding flight, force generation in flapping flight, aerodynamic efficiency, comparative flight performance and pursuit strategies during predatory and territorial flights. New data are set in context by brief reviews covering anatomy at several scales, insect aerodynamics, neuromechanics and behaviour. We achieve a new perspective by means of a diverse range of techniques, including laser-line mapping of wing topographies, computational fluid dynamics simulations of finely detailed wing geometries, quantitative imaging using particle image velocimetry of on-wing and wake flow patterns, classical aerodynamic theory, photography in the field, infrared motion capture and multi-camera optical tracking of free flight trajectories in laboratory environments. Our comprehensive approach enables a novel synthesis of datasets and subfields that integrates many aspects of flight from the neurobiology of the compound eye, through the aeromechanical interface with the surrounding fluid, to flight performance under cruising and higher-energy behavioural modes
- …