43 research outputs found
Computer Graphic and Photographic Image Classification using Local Image Descriptors
With the tremendous development of computer graphic rendering technology, photorealistic computer graphic images are difficult to differentiate from photo graphic images. In this article, a method is proposed based on discrete wavelet transform based binary statistical image features to distinguish computer graphic from photo graphic images using the support vector machine classifier. Textural descriptors extracted using binary statistical image features are different for computer graphic and photo graphic which are based on learning of natural image statistic filters. Input RGB image is first converted into grayscale and decomposed into sub-bands using Haar discrete wavelet transform and then binary statistical image features are extracted. Fuzzy entropy based feature subset selection is employed to choose relevant features. Experimental results using Columbia database show that the method achieves good detection accuracy
Modular Convolutional Neural Network for Discriminating between Computer-Generated Images and Photographic Images
International audienceDiscriminating between computer-generated images (CGIs) and photographic images (PIs) is not a new problem in digital image forensics. However, with advances in rendering techniques supported by strong hardware and in genera-tive adversarial networks, CGIs are becoming indistinguishable from PIs in both human and computer perception. This means that malicious actors can use CGIs for spoofing facial authentication systems, impersonating other people, and creating fake news to be spread on social networks. The methods developed for discriminating between CGIs and PIs quickly become outdated and must be regularly enhanced to be able to reduce these attack surfaces. Leveraging recent advances in deep convolutional networks, we have built a modular CGI-PI discriminator with a customized VGG-19 network as the feature extractor, statistical convolutional neural networks as the feature transformers, and a discriminator. We also devised a probabilistic patch aggregation strategy to deal with high-resolution images. This proposed method outper-formed a state-of-the-art method and achieved accuracy up to 100%
Joint Learning of Deep Texture and High-Frequency Features for Computer-Generated Image Detection
Distinguishing between computer-generated (CG) and natural photographic (PG)
images is of great importance to verify the authenticity and originality of
digital images. However, the recent cutting-edge generation methods enable high
qualities of synthesis in CG images, which makes this challenging task even
trickier. To address this issue, a joint learning strategy with deep texture
and high-frequency features for CG image detection is proposed. We first
formulate and deeply analyze the different acquisition processes of CG and PG
images. Based on the finding that multiple different modules in image
acquisition will lead to different sensitivity inconsistencies to the
convolutional neural network (CNN)-based rendering in images, we propose a deep
texture rendering module for texture difference enhancement and discriminative
texture representation. Specifically, the semantic segmentation map is
generated to guide the affine transformation operation, which is used to
recover the texture in different regions of the input image. Then, the
combination of the original image and the high-frequency components of the
original and rendered images are fed into a multi-branch neural network
equipped with attention mechanisms, which refines intermediate features and
facilitates trace exploration in spatial and channel dimensions respectively.
Extensive experiments on two public datasets and a newly constructed dataset
with more realistic and diverse images show that the proposed approach
outperforms existing methods in the field by a clear margin. Besides, results
also demonstrate the detection robustness and generalization ability of the
proposed approach to postprocessing operations and generative adversarial
network (GAN) generated images
Interactive real-time three-dimensional visualisation of virtual textiles
Virtual textile databases provide a cost-efficient alternative to the use of existing hardcover
sample catalogues. By taking advantage of the high performance features offered by the
latest generation of programmable graphics accelerator boards, it is possible to combine
photometric stereo methods with 3D visualisation methods to implement a virtual textile
database. In this thesis, we investigate and combine rotation invariant texture retrieval with
interactive visualisation techniques.
We use a 3D surface representation that is a generic data representation that allows us to
combine real-time interactive 3D visualisation methods with present day texture retrieval
methods. We begin by investigating the most suitable data format for the 3D surface
representation and identify relief-mapping combined with Bézier surfaces as the most
suitable 3D surface representations for our needs, and go on to describe how these
representation can be combined for real-time rendering. We then investigate ten different
methods of implementing rotation invariant texture retrieval using feature vectors. These
results show that first order statistics in the form of histogram data are very effective for
discriminating colour albedo information, while rotation invariant gradient maps are
effective for distinguishing between different types of micro-geometry using either first or
second order statistics.Engineering and physical Sciences Research (EPSRC
Self-Supervised Shape and Appearance Modeling via Neural Differentiable Graphics
Inferring 3D shape and appearance from natural images is a fundamental challenge in computer vision. Despite recent progress using deep learning methods, a key limitation is the availability of annotated training data, as acquisition is often very challenging and expensive, especially at a large scale. This thesis proposes to incorporate physical priors into neural networks that allow for self-supervised learning.
As a result, easy-to-access unlabeled data can be used for model training. In particular, novel algorithms in the context of 3D reconstruction and texture/material synthesis are introduced, where only image data is available as supervisory signal.
First, a method that learns to reason about 3D shape and appearance solely from unstructured 2D images, achieved via differentiable rendering in an adversarial fashion, is proposed.
As shown next, learning from videos significantly improves 3D reconstruction quality. To this end, a novel ray-conditioned warp embedding is proposed that aggregates pixel-wise features from multiple source images.
Addressing the challenging task of disentangling shape and appearance, first a method that enables 3D texture synthesis independent of shape or resolution is presented. For this purpose, 3D noise fields of different scales are transformed into stationary textures. The method is able to produce 3D textures, despite only requiring 2D textures for training.
Lastly, the surface characteristics of textures under different illumination conditions are modeled in the form of material parameters. Therefore, a self-supervised approach is proposed that has no access to material parameters but only flash images. Similar to the previous method, random noise fields are reshaped to material parameters, which are conditioned to replicate the visual appearance of the input under matching light
Management and display of four-dimensional environmental data sets using McIDAS
Over the past four years, great strides have been made in the areas of data management and display of 4-D meteorological data sets. A survey was conducted of available and planned 4-D meteorological data sources. The data types were evaluated for their impact on the data management and display system. The requirements were analyzed for data base management generated by the 4-D data display system. The suitability of the existing data base management procedures and file structure were evaluated in light of the new requirements. Where needed, new data base management tools and file procedures were designed and implemented. The quality of the basic 4-D data sets was assured. The interpolation and extrapolation techniques of the 4-D data were investigated. The 4-D data from various sources were combined to make a uniform and consistent data set for display purposes. Data display software was designed to create abstract line graphic 3-D displays. Realistic shaded 3-D displays were created. Animation routines for these displays were developed in order to produce a dynamic 4-D presentation. A prototype dynamic color stereo workstation was implemented. A computer functional design specification was produced based on interactive studies and user feedback
Acquisition, Modeling, and Augmentation of Reflectance for Synthetic Optical Flow Reference Data
This thesis is concerned with the acquisition, modeling, and augmentation of material reflectance to simulate high-fidelity synthetic data for computer vision tasks.
The topic is covered in three chapters: I commence with exploring the upper limits of reflectance acquisition.
I analyze state-of-the-art BTF reflectance field renderings and show that they can be applied to optical flow performance analysis with closely matching performance to real-world images.
Next, I present two methods for fitting efficient BRDF reflectance models to measured BTF data.
Both methods combined retain all relevant reflectance information as well as the surface normal details on a pixel level.
I further show that the resulting synthesized images are suited for optical flow performance analysis, with a virtually identical performance for all material types.
Finally, I present a novel method for augmenting real-world datasets with physically plausible precipitation effects, including ground surface wetting, water droplets on the windshield, and water spray and mists.
This is achieved by projecting the realworld image data onto a reconstructed virtual scene, manipulating the scene and the surface reflectance, and performing unbiased light transport simulation of the precipitation effects
Digital 3D reconstruction as a research environment in art and architecture history: uncertainty classification and visualisation
The dissertation addresses the still not solved challenges concerned with the source-based digital 3D reconstruction, visualisation and documentation in the domain of archaeology, art and architecture history.
The emerging BIM methodology and the exchange data format IFC are changing the way of collaboration, visualisation and documentation in the planning, construction and facility management process. The introduction and development of the Semantic Web (Web 3.0), spreading the idea of structured, formalised and linked data, offers semantically enriched human- and machine-readable data.
In contrast to civil engineering and cultural heritage, academic object-oriented disciplines, like archaeology, art and architecture history, are acting as outside spectators.
Since the 1990s, it has been argued that a 3D model is not likely to be considered a scientific reconstruction unless it is grounded on accurate documentation and visualisation. However, these standards are still missing and the validation of the outcomes is not fulfilled. Meanwhile, the digital research data remain ephemeral and continue to fill the growing digital cemeteries.
This study focuses, therefore, on the evaluation of the source-based digital 3D reconstructions and, especially, on uncertainty assessment in the case of hypothetical reconstructions of destroyed or never built artefacts according to scientific principles, making the models shareable and reusable by a potentially wide audience.
The work initially focuses on terminology and on the definition of a workflow especially related to the classification and visualisation of uncertainty. The workflow is then applied to specific cases of 3D models uploaded to the DFG repository of the AI Mainz. In this way, the available methods of documenting, visualising and communicating uncertainty are analysed.
In the end, this process will lead to a validation or a correction of the workflow and the initial assumptions, but also (dealing with different hypotheses) to a better definition of the levels of uncertainty