23,921 research outputs found
Alignment-free HDR Deghosting with Semantics Consistent Transformer
High dynamic range (HDR) imaging aims to retrieve information from multiple
low-dynamic range inputs to generate realistic output. The essence is to
leverage the contextual information, including both dynamic and static
semantics, for better image generation. Existing methods often focus on the
spatial misalignment across input frames caused by the foreground and/or camera
motion. However, there is no research on jointly leveraging the dynamic and
static context in a simultaneous manner. To delve into this problem, we propose
a novel alignment-free network with a Semantics Consistent Transformer (SCTNet)
with both spatial and channel attention modules in the network. The spatial
attention aims to deal with the intra-image correlation to model the dynamic
motion, while the channel attention enables the inter-image intertwining to
enhance the semantic consistency across frames. Aside from this, we introduce a
novel realistic HDR dataset with more variations in foreground objects,
environmental factors, and larger motions. Extensive comparisons on both
conventional datasets and ours validate the effectiveness of our method,
achieving the best trade-off on the performance and the computational cost
CVABS: Moving Object Segmentation with Common Vector Approach for Videos
Background modelling is a fundamental step for several real-time computer
vision applications that requires security systems and monitoring. An accurate
background model helps detecting activity of moving objects in the video. In
this work, we have developed a new subspace based background modelling
algorithm using the concept of Common Vector Approach with Gram-Schmidt
orthogonalization. Once the background model that involves the common
characteristic of different views corresponding to the same scene is acquired,
a smart foreground detection and background updating procedure is applied based
on dynamic control parameters. A variety of experiments is conducted on
different problem types related to dynamic backgrounds. Several types of
metrics are utilized as objective measures and the obtained visual results are
judged subjectively. It was observed that the proposed method stands
successfully for all problem types reported on CDNet2014 dataset by updating
the background frames with a self-learning feedback mechanism.Comment: 12 Pages, 4 Figures, 1 Tabl
Roadmap on optical security
Postprint (author's final draft
Roadmap on structured light
Structured light refers to the generation and application of custom light fields. As the tools and technology to create and detect structured light have evolved, steadily the applications have begun to emerge. This roadmap touches on the key fields within structured light from the perspective of experts in those areas, providing insight into the current state and the challenges their respective fields face. Collectively the roadmap outlines the venerable nature of structured light research and the exciting prospects for the future that are yet to be realized.Peer ReviewedPostprint (published version
Photoelastic stress analysis under unconventional loading
This paper presents use of conventional photoelastic techniques under unconventional loading situations to evaluate their efficacy in sensing applications. The loading is unconventional in the sense that low modulus photoelastic material is deformed under vertical load in the direction of light travel to induce the photoelastic effect. This is atypical of conventional methods where loading is across the light travel. Both RGB calibration and phase shining techniques have been used to study the characteristics of fringe patterns obtained under vertical and shear loads. The results obtained under these conditions are discussed with their limitations specially when this is applied for sensing applications. Finally a case study has been conducted to analyze the foot image and conclusions drawn from this have been presented. Copyright © 2007 by ASME
LAN-HDR: Luminance-based Alignment Network for High Dynamic Range Video Reconstruction
As demands for high-quality videos continue to rise, high-resolution and
high-dynamic range (HDR) imaging techniques are drawing attention. To generate
an HDR video from low dynamic range (LDR) images, one of the critical steps is
the motion compensation between LDR frames, for which most existing works
employed the optical flow algorithm. However, these methods suffer from flow
estimation errors when saturation or complicated motions exist. In this paper,
we propose an end-to-end HDR video composition framework, which aligns LDR
frames in the feature space and then merges aligned features into an HDR frame,
without relying on pixel-domain optical flow. Specifically, we propose a
luminance-based alignment network for HDR (LAN-HDR) consisting of an alignment
module and a hallucination module. The alignment module aligns a frame to the
adjacent reference by evaluating luminance-based attention, excluding color
information. The hallucination module generates sharp details, especially for
washed-out areas due to saturation. The aligned and hallucinated features are
then blended adaptively to complement each other. Finally, we merge the
features to generate a final HDR frame. In training, we adopt a temporal loss,
in addition to frame reconstruction losses, to enhance temporal consistency and
thus reduce flickering. Extensive experiments demonstrate that our method
performs better or comparable to state-of-the-art methods on several
benchmarks.Comment: ICCV 202
- …