746 research outputs found
Looking at faces in the wild
Recent advances in the face detection (FD) and recognition (FR) technology may give an impression that the problem of face matching is essentially solved, e.g. via deep learning models using thousands of samples per face for training and validation on the available benchmark data-sets. Human vision system seems to handle face localization and matching problem differently from the modern FR systems, since humans detect faces instantly even in most cluttered environments, and often require a single view of a face to reliably distinguish it from all others. This prompted us to take a biologically inspired look at building a cognitive architecture that uses artificial neural nets at the face detection stage and adapts a single image per person (SIPP) approach for face image matching
cGAN-based Manga Colorization Using a Single Training Image
The Japanese comic format known as Manga is popular all over the world. It is
traditionally produced in black and white, and colorization is time consuming
and costly. Automatic colorization methods generally rely on greyscale values,
which are not present in manga. Furthermore, due to copyright protection,
colorized manga available for training is scarce. We propose a manga
colorization method based on conditional Generative Adversarial Networks
(cGAN). Unlike previous cGAN approaches that use many hundreds or thousands of
training images, our method requires only a single colorized reference image
for training, avoiding the need of a large dataset. Colorizing manga using
cGANs can produce blurry results with artifacts, and the resolution is limited.
We therefore also propose a method of segmentation and color-correction to
mitigate these issues. The final results are sharp, clear, and in high
resolution, and stay true to the character's original color scheme.Comment: 8 pages, 13 figure
Our Deep CNN Face Matchers Have Developed Achromatopsia
Modern deep CNN face matchers are trained on datasets containing color
images. We show that such matchers achieve essentially the same accuracy on the
grayscale or the color version of a set of test images. We then consider
possible causes for deep CNN face matchers ``not seeing color''. Popular
web-scraped face datasets actually have 30 to 60\% of their identities with one
or more grayscale images. We analyze whether this grayscale element in the
training set impacts the accuracy achieved, and conclude that it does not.
Further, we show that even with a 100\% grayscale training set, comparable
accuracy is achieved on color or grayscale test images. Then we show that the
skin region of an individual's images in a web-scraped training set exhibit
significant variation in their mapping to color space. This suggests that
color, at least for web-scraped, in-the-wild face datasets, carries limited
identity-related information for training state-of-the-art matchers. Finally,
we verify that comparable accuracy is achieved from training using
single-channel grayscale images, implying that a larger dataset can be used
within the same memory limit, with a less computationally intensive early
layer
Inharmonious Region Localization by Magnifying Domain Discrepancy
Inharmonious region localization aims to localize the region in a synthetic
image which is incompatible with surrounding background. The inharmony issue is
mainly attributed to the color and illumination inconsistency produced by image
editing techniques. In this work, we tend to transform the input image to
another color space to magnify the domain discrepancy between inharmonious
region and background, so that the model can identify the inharmonious region
more easily. To this end, we present a novel framework consisting of a color
mapping module and an inharmonious region localization network, in which the
former is equipped with a novel domain discrepancy magnification loss and the
latter could be an arbitrary localization network. Extensive experiments on
image harmonization dataset show the superiority of our designed framework. Our
code is available at
https://github.com/bcmi/MadisNet-Inharmonious-Region-Localization
Perceiving Unknown in Dark from Perspective of Cell Vibration
Low light very likely leads to the degradation of image quality and even
causes visual tasks' failure. Existing image enhancement technologies are prone
to over-enhancement or color distortion, and their adaptability is fairly
limited. In order to deal with these problems, we utilise the mechanism of
biological cell vibration to interpret the formation of color images. In
particular, we here propose a simple yet effective cell vibration energy (CVE)
mapping method for image enhancement. Based on a hypothetical color-formation
mechanism, our proposed method first uses cell vibration and photoreceptor
correction to determine the photon flow energy for each color channel, and then
reconstructs the color image with the maximum energy constraint of the visual
system. Photoreceptor cells can adaptively adjust the feedback from the light
intensity of the perceived environment. Based on this understanding, we here
propose a new Gamma auto-adjustment method to modify Gamma values according to
individual images. Finally, a fusion method, combining CVE and Gamma
auto-adjustment (CVE-G), is proposed to reconstruct the color image under the
constraint of lightness. Experimental results show that the proposed algorithm
is superior to six state of the art methods in avoiding over-enhancement and
color distortion, restoring the textures of dark areas and reproducing natural
colors. The source code will be released at
https://github.com/leixiaozhou/CVE-G-Resource-Base.Comment: 13 pages, 17 figure
- …