491 research outputs found

    IML-ViT: Benchmarking Image Manipulation Localization by Vision Transformer

    Full text link
    Advanced image tampering techniques are increasingly challenging the trustworthiness of multimedia, leading to the development of Image Manipulation Localization (IML). But what makes a good IML model? The answer lies in the way to capture artifacts. Exploiting artifacts requires the model to extract non-semantic discrepancies between manipulated and authentic regions, necessitating explicit comparisons between the two areas. With the self-attention mechanism, naturally, the Transformer should be a better candidate to capture artifacts. However, due to limited datasets, there is currently no pure ViT-based approach for IML to serve as a benchmark, and CNNs dominate the entire task. Nevertheless, CNNs suffer from weak long-range and non-semantic modeling. To bridge this gap, based on the fact that artifacts are sensitive to image resolution, amplified under multi-scale features, and massive at the manipulation border, we formulate the answer to the former question as building a ViT with high-resolution capacity, multi-scale feature extraction capability, and manipulation edge supervision that could converge with a small amount of data. We term this simple but effective ViT paradigm IML-ViT, which has significant potential to become a new benchmark for IML. Extensive experiments on five benchmark datasets verified our model outperforms the state-of-the-art manipulation localization methods.Code and models are available at \url{https://github.com/SunnyHaze/IML-ViT}

    Local blur estimation based on toggle mapping

    No full text
    International audienceA local blur estimation method is proposed, based on the difference between the gradient and the residue of the toggle mapping. This method is able to compare the quality of images with different content and does not require a contour detection step. Qualitative results are shown in the context of the LINX project. Then, quantitative results are given on DIQA database, outperforming the combination of classical blur detection methods reported in the literature

    A Forensic Scheme for Revealing Post-processed Region Duplication Forgery in Suspected Images

    Get PDF
    Recent researches have demonstrated that local interest points alone can be employed to detect region duplication forgery in image forensics. Authentic images may be abused by copy-move tool in Adobe Photoshop to fully contained duplicated regions such as objects with high primitives such as corners and edges. Corners and edges represent the internal structure of an object in the image which makes them have a discriminating property under geometric transformations such as scale and rotation operation. They can be localised using scale-invariant features transform (SIFT) algorithm. In this paper, we provide an image forgery detection technique by using local interest points. Local interest points can be exposed by extracting adaptive non-maximal suppression (ANMS) keypoints from dividing blocks in the segmented image to detect such corners of objects. We also demonstrate that ANMS keypoints can be effectively utilised to detect blurred and scaled forged regions. The ANMS features of the image are shown to exhibit the internal structure of copy moved region. We provide a new texture descriptor called local phase quantisation (LPQ) that is robust to image blurring and also to eliminate the false positives of duplicated regions. Experimental results show that our scheme has the ability to reveal region duplication forgeries under scaling, rotation and blur manipulation of JPEG images on MICC-F220 and CASIA v2 image datasets

    Tamper detection of qur'anic text watermarking scheme based on vowel letters with Kashida using exclusive-or and queueing technique

    Get PDF
    The most sensitive Arabic text available online is the digital Holy Qur’an. This sacred Islamic religious book is recited by all Muslims worldwide including the non-Arabs as part of their worship needs. It should be protected from any kind of tampering to keep its invaluable meaning intact. Different characteristics of the Arabic letters like the vowels ( أ . و . ي ), Kashida (extended letters), and other symbols in the Holy Qur’an must be secured from alterations. The cover text of the al-Qur’an and its watermarked text are different due to the low values of the Peak Signal to Noise Ratio (PSNR), Embedding Ratio (ER), and Normalized Cross-Correlation (NCC), thus the location for tamper detection gets low accuracy. Watermarking technique with enhanced attributes must therefore be designed for the Qur’an text using Arabic vowel letters with Kashida. Most of the existing detection methods that tried to achieve accurate results related to the tampered Qur’an text often show various limitations like diacritics, alif mad surah, double space, separate shapes of Arabic letters, and Kashida. The gap addressed by this research is to improve the security of Arabic text in the Holy Qur’an by using vowel letters with Kashida. The purpose of this research is to enhance Quran text watermarking scheme based on exclusive-or and reversing with queueing techniques. The methodology consists of four phases. The first phase is pre-processing followed by the embedding process phase to hide the data after the vowel letters wherein if the secret bit is ‘1’, insert the Kashida but do not insert it if the bit is ‘0’. The third phase is extraction process and the last phase is to evaluate the performance of the proposed scheme by using PSNR (for the imperceptibility), ER (for the capacity), and NCC (for the security of the watermarking). The experimental results revealed the improvement of the NCC by 1.77 %, PSNR by 9.6 %, and ER by 8.6 % compared to available current schemes. Hence, it can be concluded that the proposed scheme has the ability to detect the location of tampering accurately for attacks of insertion, deletion, and reordering
    corecore