7 research outputs found
Harmonizer: Learning to Perform White-Box Image and Video Harmonization
Recent works on image harmonization solve the problem as a pixel-wise image
translation task via large autoencoders. They have unsatisfactory performances
and slow inference speeds when dealing with high-resolution images. In this
work, we observe that adjusting the input arguments of basic image filters,
e.g., brightness and contrast, is sufficient for humans to produce realistic
images from the composite ones. Hence, we frame image harmonization as an
image-level regression problem to learn the arguments of the filters that
humans use for the task. We present a Harmonizer framework for image
harmonization. Unlike prior methods that are based on black-box autoencoders,
Harmonizer contains a neural network for filter argument prediction and several
white-box filters (based on the predicted arguments) for image harmonization.
We also introduce a cascade regressor and a dynamic loss strategy for
Harmonizer to learn filter arguments more stably and precisely. Since our
network only outputs image-level arguments and the filters we used are
efficient, Harmonizer is much lighter and faster than existing methods.
Comprehensive experiments demonstrate that Harmonizer surpasses existing
methods notably, especially with high-resolution inputs. Finally, we apply
Harmonizer to video harmonization, which achieves consistent results across
frames and 56 fps at 1080P resolution. Code and models are available at:
https://github.com/ZHKKKe/Harmonizer
Neural Preset for Color Style Transfer
In this paper, we present a Neural Preset technique to address the
limitations of existing color style transfer methods, including visual
artifacts, vast memory requirement, and slow style switching speed. Our method
is based on two core designs. First, we propose Deterministic Neural Color
Mapping (DNCM) to consistently operate on each pixel via an image-adaptive
color mapping matrix, avoiding artifacts and supporting high-resolution inputs
with a small memory footprint. Second, we develop a two-stage pipeline by
dividing the task into color normalization and stylization, which allows
efficient style switching by extracting color styles as presets and reusing
them on normalized input images. Due to the unavailability of pairwise
datasets, we describe how to train Neural Preset via a self-supervised
strategy. Various advantages of Neural Preset over existing methods are
demonstrated through comprehensive evaluations. Notably, Neural Preset enables
stable 4K color style transfer in real-time without artifacts. Besides, we show
that our trained model can naturally support multiple applications without
fine-tuning, including low-light image enhancement, underwater image
correction, image dehazing, and image harmonization. Project page with demos:
https://zhkkke.github.io/NeuralPreset .Comment: Project page with demos: https://zhkkke.github.io/NeuralPreset .
Artifact-free real-time 4K color style transfer via AI-generated presets.
CVPR 202
Structure-Informed Shadow Removal Networks
Existing deep learning-based shadow removal methods still produce images with
shadow remnants. These shadow remnants typically exist in homogeneous regions
with low-intensity values, making them untraceable in the existing
image-to-image mapping paradigm. We observe that shadows mainly degrade images
at the image-structure level (in which humans perceive object shapes and
continuous colors). Hence, in this paper, we propose to remove shadows at the
image structure level. Based on this idea, we propose a novel
structure-informed shadow removal network (StructNet) to leverage the
image-structure information to address the shadow remnant problem.
Specifically, StructNet first reconstructs the structure information of the
input image without shadows and then uses the restored shadow-free structure
prior to guiding the image-level shadow removal. StructNet contains two main
novel modules: (1) a mask-guided shadow-free extraction (MSFE) module to
extract image structural features in a non-shadow-to-shadow directional manner,
and (2) a multi-scale feature & residual aggregation (MFRA) module to leverage
the shadow-free structure information to regularize feature consistency. In
addition, we also propose to extend StructNet to exploit multi-level structure
information (MStructNet), to further boost the shadow removal performance with
minimum computational overheads. Extensive experiments on three shadow removal
benchmarks demonstrate that our method outperforms existing shadow removal
methods, and our StructNet can be integrated with existing methods to improve
them further.Comment: IEEE TI