16 research outputs found
Real-World Blur Dataset for Learning and Benchmarking Deblurring Algorithms
Computational Photography, Deblurring, Low-level Vision, Datasets and EvaluationNumerous learning-based approaches to single image deblurring for camera and object motion blurs have recently been proposed. To generalize such approaches to real-world blurs, large datasets of real blurred images and their ground truth sharp images are essential. However, there are still no such datasets, thus all the existing approaches resort to synthetic ones, which leads to the failure of deblurring real-world images. In this work, we present a large-scale dataset of real-world blurred images and their corresponding sharp images captured in low-light environments for learning and benchmarking single image deblurring methods. To collect our dataset, we build an image acquisition system to simultaneously capture a geometrically aligned pair of blurred and sharp images, and develop a post-processing method to further align images geometrically and photometrically. We analyze the effect of our post-processing step, and the performance of existing learning-based deblurring methods. Our analysis shows that our dataset significantly improves deblurring quality for real-world low-light images.Y1. Introduction 1
2. Related Work 2
3. Image Acquisition System and Process 3
3.1 Image Acquisition System 3
3.2 Image Acquisition Process 4
4. Post-Processing 5
4.1 Downsampling & Denoising 6
4.2 Geometric Alignment 6
4.3 Photometric Alignment 8
5. Experiments 8
5.1 Analysis of RealBlur Dataset 9
5.2 Benchmark 12
6. Conclusion 19
7. Appendix 20
8. References 24
9. 요약문 28MasterdCollectio
Estimates of M-Harmonic Conjugate Operator
We define the M-harmonic conjugate operator K and prove that for 1<p<∞, there is a constant Cp such that ∫S|Kf|pωdσ≤Cp∫S|f|pωdσ for all f∈Lp(ω) if and only if the nonnegative weight ω satisfies the Ap-condition. Also, we prove that if there is a constant Cp such that ∫S|Kf|pvdσ≤Cp∫S|f|pwdσ for all f∈Lp(w), then the pair of weights (v,w) satisfies the Ap-condition
Estimates of weighted Hardy–Littlewood averages on the p-adic vector space
AbstractIn the p-adic vector space Qpn, we characterize those non-negative functions ψ defined on Zp*={w∈Qp:0<|w|p⩽1} for which the weighted Hardy–Littlewood average Uψ:f→∫Zp*f(t⋅)ψ(t)dt is bounded on Lr(Qpn) (1⩽r⩽∞), and on BMO(Qpn). Also, in each case, we find the corresponding operator norm ‖Uψ‖
ExBluRF: Efficient Radiance Fields for Extreme Motion Blurred Images
We present ExBluRF, a novel view synthesis method for extreme motion blurred
images based on efficient radiance fields optimization. Our approach consists
of two main components: 6-DOF camera trajectory-based motion blur formulation
and voxel-based radiance fields. From extremely blurred images, we optimize the
sharp radiance fields by jointly estimating the camera trajectories that
generate the blurry images. In training, multiple rays along the camera
trajectory are accumulated to reconstruct single blurry color, which is
equivalent to the physical motion blur operation. We minimize the
photo-consistency loss on blurred image space and obtain the sharp radiance
fields with camera trajectories that explain the blur of all images. The joint
optimization on the blurred image space demands painfully increasing
computation and resources proportional to the blur size. Our method solves this
problem by replacing the MLP-based framework to low-dimensional 6-DOF camera
poses and voxel-based radiance fields. Compared with the existing works, our
approach restores much sharper 3D scenes from challenging motion blurred views
with the order of 10 times less training time and GPU memory consumption.Comment: https://github.com/taekkii/ExBluRF/tree/mai
Human Pose Estimation in Extremely Low-Light Conditions
We study human pose estimation in extremely low-light images. This task is
challenging due to the difficulty of collecting real low-light images with
accurate labels, and severely corrupted inputs that degrade prediction quality
significantly. To address the first issue, we develop a dedicated camera system
and build a new dataset of real low-light images with accurate pose labels.
Thanks to our camera system, each low-light image in our dataset is coupled
with an aligned well-lit image, which enables accurate pose labeling and is
used as privileged information during training. We also propose a new model and
a new training strategy that fully exploit the privileged information to learn
representation insensitive to lighting conditions. Our method demonstrates
outstanding performance on real extremely low light images, and extensive
analyses validate that both of our model and dataset contribute to the success.Comment: Accepted to CVPR 202
Estimates of <inline-formula> <graphic file="1029-242X-2010-435450-i1.gif"/></inline-formula>-Harmonic Conjugate Operator
We define the -harmonic conjugate operator and prove that for , there is a constant such that for all if and only if the nonnegative weight satisfies the -condition. Also, we prove that if there is a constant such that for all , then the pair of weights satisfies the -condition.</p
Iterative Filter Adaptive Network for Single Image Defocus Deblurring
We propose a novel end-to-end learning-based approach for single image defocus deblurring. The proposed approach is equipped with a novel Iterative Filter Adaptive Network (IFAN) that is specifically designed to handle spatially-varying and large defocus blur. For adaptively handling spatially-varying blur, IFAN predicts pixel-wise deblurring filters, which are applied to defocused features of an input image to generate deblurred features. For effectively managing large blur, IFAN models deblurring filters as stacks of small-sized separable filters. Predicted separable deblurring filters are applied to defocused features using a novel Iterative Adaptive Convolution (IAC) layer. We also propose a training scheme based on defocus disparity estimation and reblurring, which significantly boosts the deblurring quality. We demonstrate that our method achieves state-of-the-art performance both quantitatively and qualitatively on real-world images1