1,147 research outputs found
Learning Rich Features for Image Manipulation Detection
Image manipulation detection is different from traditional semantic object
detection because it pays more attention to tampering artifacts than to image
content, which suggests that richer features need to be learned. We propose a
two-stream Faster R-CNN network and train it endto- end to detect the tampered
regions given a manipulated image. One of the two streams is an RGB stream
whose purpose is to extract features from the RGB image input to find tampering
artifacts like strong contrast difference, unnatural tampered boundaries, and
so on. The other is a noise stream that leverages the noise features extracted
from a steganalysis rich model filter layer to discover the noise inconsistency
between authentic and tampered regions. We then fuse features from the two
streams through a bilinear pooling layer to further incorporate spatial
co-occurrence of these two modalities. Experiments on four standard image
manipulation datasets demonstrate that our two-stream framework outperforms
each individual stream, and also achieves state-of-the-art performance compared
to alternative methods with robustness to resizing and compression.Comment: CVPR 2018 Camera Read
Autoencoder with recurrent neural networks for video forgery detection
Video forgery detection is becoming an important issue in recent years,
because modern editing software provide powerful and easy-to-use tools to
manipulate videos. In this paper we propose to perform detection by means of
deep learning, with an architecture based on autoencoders and recurrent neural
networks. A training phase on a few pristine frames allows the autoencoder to
learn an intrinsic model of the source. Then, forged material is singled out as
anomalous, as it does not fit the learned model, and is encoded with a large
reconstruction error. Recursive networks, implemented with the long short-term
memory model, are used to exploit temporal dependencies. Preliminary results on
forged videos show the potential of this approach.Comment: Presented at IS&T Electronic Imaging: Media Watermarking, Security,
and Forensics, January 201
D-Unet: A Dual-encoder U-Net for Image Splicing Forgery Detection and Localization
Recently, many detection methods based on convolutional neural networks
(CNNs) have been proposed for image splicing forgery detection. Most of these
detection methods focus on the local patches or local objects. In fact, image
splicing forgery detection is a global binary classification task that
distinguishes the tampered and non-tampered regions by image fingerprints.
However, some specific image contents are hardly retained by CNN-based
detection networks, but if included, would improve the detection accuracy of
the networks. To resolve these issues, we propose a novel network called
dual-encoder U-Net (D-Unet) for image splicing forgery detection, which employs
an unfixed encoder and a fixed encoder. The unfixed encoder autonomously learns
the image fingerprints that differentiate between the tampered and non-tampered
regions, whereas the fixed encoder intentionally provides the direction
information that assists the learning and detection of the network. This
dual-encoder is followed by a spatial pyramid global-feature extraction module
that expands the global insight of D-Unet for classifying the tampered and
non-tampered regions more accurately. In an experimental comparison study of
D-Unet and state-of-the-art methods, D-Unet outperformed the other methods in
image-level and pixel-level detection, without requiring pre-training or
training on a large number of forgery images. Moreover, it was stably robust to
different attacks.Comment: 13 pages, 13 figure
- …