427 research outputs found
Autoencoder with recurrent neural networks for video forgery detection
Video forgery detection is becoming an important issue in recent years,
because modern editing software provide powerful and easy-to-use tools to
manipulate videos. In this paper we propose to perform detection by means of
deep learning, with an architecture based on autoencoders and recurrent neural
networks. A training phase on a few pristine frames allows the autoencoder to
learn an intrinsic model of the source. Then, forged material is singled out as
anomalous, as it does not fit the learned model, and is encoded with a large
reconstruction error. Recursive networks, implemented with the long short-term
memory model, are used to exploit temporal dependencies. Preliminary results on
forged videos show the potential of this approach.Comment: Presented at IS&T Electronic Imaging: Media Watermarking, Security,
and Forensics, January 201
A PatchMatch-based Dense-field Algorithm for Video Copy-Move Detection and Localization
We propose a new algorithm for the reliable detection and localization of
video copy-move forgeries. Discovering well crafted video copy-moves may be
very difficult, especially when some uniform background is copied to occlude
foreground objects. To reliably detect both additive and occlusive copy-moves
we use a dense-field approach, with invariant features that guarantee
robustness to several post-processing operations. To limit complexity, a
suitable video-oriented version of PatchMatch is used, with a multiresolution
search strategy, and a focus on volumes of interest. Performance assessment
relies on a new dataset, designed ad hoc, with realistic copy-moves and a wide
variety of challenging situations. Experimental results show the proposed
method to detect and localize video copy-moves with good accuracy even in
adverse conditions
Information Forensics and Security: A quarter-century-long journey
Information forensics and security (IFS) is an active R&D area whose goal is to ensure that people use devices, data, and intellectual properties for authorized purposes and to facilitate the gathering of solid evidence to hold perpetrators accountable. For over a quarter century, since the 1990s, the IFS research area has grown tremendously to address the societal needs of the digital information era. The IEEE Signal Processing Society (SPS) has emerged as an important hub and leader in this area, and this article celebrates some landmark technical contributions. In particular, we highlight the major technological advances by the research community in some selected focus areas in the field during the past 25 years and present future trends
A Modified Fourier-Mellin Approach for Source Device Identification on Stabilized Videos
To decide whether a digital video has been captured by a given device,
multimedia forensic tools usually exploit characteristic noise traces left by
the camera sensor on the acquired frames. This analysis requires that the noise
pattern characterizing the camera and the noise pattern extracted from video
frames under analysis are geometrically aligned. However, in many practical
scenarios this does not occur, thus a re-alignment or synchronization has to be
performed. Current solutions often require time consuming search of the
realignment transformation parameters. In this paper, we propose to overcome
this limitation by searching scaling and rotation parameters in the frequency
domain. The proposed algorithm tested on real videos from a well-known
state-of-the-art dataset shows promising results
Transdisciplinary AI Observatory -- Retrospective Analyses and Future-Oriented Contradistinctions
In the last years, AI safety gained international recognition in the light of
heterogeneous safety-critical and ethical issues that risk overshadowing the
broad beneficial impacts of AI. In this context, the implementation of AI
observatory endeavors represents one key research direction. This paper
motivates the need for an inherently transdisciplinary AI observatory approach
integrating diverse retrospective and counterfactual views. We delineate aims
and limitations while providing hands-on-advice utilizing concrete practical
examples. Distinguishing between unintentionally and intentionally triggered AI
risks with diverse socio-psycho-technological impacts, we exemplify a
retrospective descriptive analysis followed by a retrospective counterfactual
risk analysis. Building on these AI observatory tools, we present near-term
transdisciplinary guidelines for AI safety. As further contribution, we discuss
differentiated and tailored long-term directions through the lens of two
disparate modern AI safety paradigms. For simplicity, we refer to these two
different paradigms with the terms artificial stupidity (AS) and eternal
creativity (EC) respectively. While both AS and EC acknowledge the need for a
hybrid cognitive-affective approach to AI safety and overlap with regard to
many short-term considerations, they differ fundamentally in the nature of
multiple envisaged long-term solution patterns. By compiling relevant
underlying contradistinctions, we aim to provide future-oriented incentives for
constructive dialectics in practical and theoretical AI safety research
- …