Nothing Special   »   [go: up one dir, main page]

You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (109)

Search Parameters:
Keywords = image forgery

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
19 pages, 59781 KiB  
Article
A Watermark-Based Scheme for Authenticating JPEG 2000 Image Integrity That Complies with JPEG Privacy and Security
by Jinhee Lee, Oh-Jin Kwon, Yaseen and Seungcheol Choi
Appl. Sci. 2024, 14(18), 8428; https://doi.org/10.3390/app14188428 - 19 Sep 2024
Viewed by 342
Abstract
Network development has made it easier to access multimedia material and to change it by allowing the collection, modification, and transmission of digital data. Additionally, this has led to a rise in malicious use, including unauthorized data distribution and copying. As the quantity [...] Read more.
Network development has made it easier to access multimedia material and to change it by allowing the collection, modification, and transmission of digital data. Additionally, this has led to a rise in malicious use, including unauthorized data distribution and copying. As the quantity of evil activities increases, security issues such as unauthorized use and image forgery are rising. While security solutions for JPEG-1 images are widely available, there remains a significant gap in protection for JPEG 2000 images. In this paper, we propose a new watermark-based forgery detection method to comply with the JPEG Privacy and Security standards and to authenticate JPEG 2000 image integrity in the discrete wavelet transform (DWT) domain. The method proposed divides JPEG 2000 images into groups of blocks (GOBs) within the DWT domain. The watermark is generated by collaborating with the neighboring GOBs and is embedded in the GOBs. It enables you to respond to the collage attack. Experimental results using various sample images show the superiority of the proposed method, showing negligible visual differences between the original and watermarked JPEG 2000 images. Full article
Show Figures

Figure 1

Figure 1
<p>Example of the encapsulating procedure using a watermark-based method.</p>
Full article ">Figure 2
<p>A JUMBF box generated by the encapsulating procedure.</p>
Full article ">Figure 3
<p>Basic encoding structure of JPEG 2000 image coding.</p>
Full article ">Figure 4
<p>Structure of the four-level DWT.</p>
Full article ">Figure 5
<p>Example of division using HH1 sub-band (<math display="inline"><semantics> <mrow> <mi>T</mi> <mi>b</mi> </mrow> </semantics></math> = 15, rate control factor = 10, <span class="html-italic">m</span> = 4).</p>
Full article ">Figure 6
<p>Sample images that were used for the experiment. It is not explicitly shown in this paper, but Image 1 (<math display="inline"><semantics> <mrow> <mn>512</mn> <mo>×</mo> <mn>512</mn> </mrow> </semantics></math>) used for experimental purposes is the well-known ’Lena’ image. (<b>a</b>) Image 2 (<math display="inline"><semantics> <mrow> <mn>512</mn> <mo>×</mo> <mn>512</mn> </mrow> </semantics></math>). (<b>b</b>) Image 3 (<math display="inline"><semantics> <mrow> <mn>696</mn> <mo>×</mo> <mn>568</mn> </mrow> </semantics></math>) (<b>c</b>). Image 4 (<math display="inline"><semantics> <mrow> <mn>664</mn> <mo>×</mo> <mn>560</mn> </mrow> </semantics></math>) (<b>d</b>). Image 5 (<math display="inline"><semantics> <mrow> <mn>512</mn> <mo>×</mo> <mn>472</mn> </mrow> </semantics></math>) (<b>e</b>). Image 6 (<math display="inline"><semantics> <mrow> <mn>4312</mn> <mo>×</mo> <mn>1868</mn> </mrow> </semantics></math>). (<b>f</b>) Image 7 (<math display="inline"><semantics> <mrow> <mn>1000</mn> <mo>×</mo> <mn>640</mn> </mrow> </semantics></math>). (<b>g</b>) Image 8 (<math display="inline"><semantics> <mrow> <mn>1920</mn> <mo>×</mo> <mn>1080</mn> </mrow> </semantics></math>). (<b>h</b>) Image 9 (<math display="inline"><semantics> <mrow> <mn>1000</mn> <mo>×</mo> <mn>640</mn> </mrow> </semantics></math>). (<b>i</b>) Image 10 (<math display="inline"><semantics> <mrow> <mn>4288</mn> <mo>×</mo> <mn>2848</mn> </mrow> </semantics></math>).</p>
Full article ">Figure 7
<p>Example of division. (<b>a</b>) Example of Y-component division when using level 1 HH band (rate control factor = 10, <span class="html-italic">m</span> = 4). (<b>b</b>) Example of Y-component division when using level 1, 2 HH band (rate control factor = 10, <span class="html-italic">m</span> = 1) (<b>c</b>) Example of Y-component division when using level 1, 2 HH band (rate control factor = 10, <span class="html-italic">m</span> = 4). (<b>d</b>) Example of Y-component division when using level 1 HL, LH, HH band (rate control factor = 10, <span class="html-italic">m</span> = 4). (<b>e</b>) Example of Cb-component division when using level 1 HH band (rate control factor = 10, <span class="html-italic">m</span> = 1). (<b>f</b>) Example of Cr-component division when using level 1 HH band (rate control factor = 10, <span class="html-italic">m</span> = 1). (<b>g</b>) Example of Cb-component division when using level 1, 2 HH band (rate control factor = 10, <span class="html-italic">m</span> = 1). (<b>h</b>) Example of Cr-component division when using level 1, 2 HH band (rate control factor = 10, <span class="html-italic">m</span> = 1). (<b>i</b>) Example of Cb-component division when using level 1 HL, LH, HH band (rate control factor = 10, <span class="html-italic">m</span> = 1). (<b>j</b>) Example of Cr-component division when using Level 1 HL, LH, HH band (rate control factor = 10, <span class="html-italic">m</span> = 1).</p>
Full article ">Figure 8
<p>Proposed watermark embedding scheme.</p>
Full article ">Figure 9
<p>Proposed watermark detection scheme.</p>
Full article ">Figure 10
<p>Example of an integrity-checking procedure using a watermark-based method without an authorized agency.</p>
Full article ">Figure 11
<p>Comparing PSNR between watermarked and original images.</p>
Full article ">
18 pages, 13182 KiB  
Article
Hierarchical Progressive Image Forgery Detection and Localization Method Based on UNet
by Yang Liu, Xiaofei Li, Jun Zhang, Shuohao Li, Shengze Hu and Jun Lei
Big Data Cogn. Comput. 2024, 8(9), 119; https://doi.org/10.3390/bdcc8090119 - 10 Sep 2024
Viewed by 659
Abstract
The rapid development of generative technologies has made the production of forged products easier, and AI-generated forged images are increasingly difficult to accurately detect, posing serious privacy risks and cognitive obstacles to individuals and society. Therefore, constructing an effective method that can accurately [...] Read more.
The rapid development of generative technologies has made the production of forged products easier, and AI-generated forged images are increasingly difficult to accurately detect, posing serious privacy risks and cognitive obstacles to individuals and society. Therefore, constructing an effective method that can accurately detect and locate forged regions has become an important task. This paper proposes a hierarchical and progressive forged image detection and localization method called HPUNet. This method assigns more reasonable hierarchical multi-level labels to the dataset as supervisory information at different levels, following cognitive laws. Secondly, multiple types of features are extracted from AI-generated images for detection and localization, and the detection and localization results are combined to enhance the task-relevant features. Subsequently, HPUNet expands the obtained image features into four different resolutions and performs detection and localization at different levels in a coarse-to-fine cognitive order. To address the limited feature field of view caused by inconsistent forgery sizes, we employ three sets of densely cross-connected hierarchical networks for sufficient interaction between feature images at different resolutions. Finally, a UNet network with a soft-threshold-constrained feature enhancement module is used to achieve detection and localization at different scales, and the reliance on a progressive mechanism establishes relationships between different branches. We use ACC and F1 as evaluation metrics, and extensive experiments on our method and the baseline methods demonstrate the effectiveness of our approach. Full article
Show Figures

Figure 1

Figure 1
<p>Description of forged image detection. (<b>a</b>) Classified t-SNE images of the dataset in the ResNet50 network. (<b>b</b>) Examples of AI tampering with images. (<b>c</b>) Schematic diagram of image multi-level label division.</p>
Full article ">Figure 2
<p>General structure of the HPUNet network. It combines multiple types of image features for detection and localization, and the dual-branch attention mechanism amplifies strongly relevant features while suppressing weakly relevant features. Combined with UNet to construct a hierarchical network, it achieves accurate detection and localization of forged images in a coarse-to-fine cognitive order.</p>
Full article ">Figure 3
<p>Two-branch attention fusion module.</p>
Full article ">Figure 4
<p>Diagram of feature fusion for branch <math display="inline"><semantics> <msub> <mi>θ</mi> <mn>2</mn> </msub> </semantics></math>.</p>
Full article ">Figure 5
<p>Soft-threshold dual-attention module.</p>
Full article ">Figure 6
<p>t-SNE visual comparison.</p>
Full article ">Figure 7
<p>Comparison picture of HPUNet and DA-HFNet.</p>
Full article ">Figure 8
<p>Comparison of large-scale fake image localization results.</p>
Full article ">Figure 9
<p>Comparison of small-scale fake image localization results.</p>
Full article ">
14 pages, 7087 KiB  
Article
Generated or Not Generated (GNG): The Importance of Background in the Detection of Fake Images
by Marco Tanfoni, Elia Giuseppe Ceroni, Sara Marziali, Niccolò Pancino, Marco Maggini and Monica Bianchini
Electronics 2024, 13(16), 3161; https://doi.org/10.3390/electronics13163161 - 10 Aug 2024
Viewed by 619
Abstract
Facial biometrics are widely used to reliably and conveniently recognize people in photos, in videos, or from real-time webcam streams. It is therefore of fundamental importance to detect synthetic faces in images in order to reduce the vulnerability of biometrics-based security systems. Furthermore, [...] Read more.
Facial biometrics are widely used to reliably and conveniently recognize people in photos, in videos, or from real-time webcam streams. It is therefore of fundamental importance to detect synthetic faces in images in order to reduce the vulnerability of biometrics-based security systems. Furthermore, manipulated images of faces can be intentionally shared on social media to spread fake news related to the targeted individual. This paper shows how fake face recognition models may mainly rely on the information contained in the background when dealing with generated faces, thus reducing their effectiveness. Specifically, a classifier is trained to separate fake images from real ones, using their representation in a latent space. Subsequently, the faces are segmented and the background removed, and the detection procedure is performed again, observing a significant drop in classification accuracy. Finally, an explainability tool (SHAP) is used to highlight the salient areas of the image, showing that the background and face contours crucially influence the classifier decision. Full article
(This article belongs to the Special Issue Deep Learning Approach for Secure and Trustworthy Biometric System)
Show Figures

Figure 1

Figure 1
<p>Examples of segmentation masks in Mut1ny dataset. In the experiments, the segmentation module was only used to finely remove the background in the images. To ensure a good variety of training samples, the facial images are drawn from different ethnicities, ages and genders, which are randomly rotated and with a wide facial poses angle range (from –90 to 90 degrees).</p>
Full article ">Figure 2
<p>The datasets used for the <span class="html-italic">Real</span> class samples. (<b>a</b>) reports some examples from CelebA-HQ dataset, which provides high-quality images of celebrity faces. Images are retrieved from the official TensorFlow dataset documentation page (accessed on 3 August 2024). (<b>b</b>) is a teaser figure for the Flickr-Faces-HQ dataset, which offers a diverse set of human face images sourced from Flickr. Images are retrieved from the NVlabs ffhq-dataset GitHub repository (accessed on 3 August 2024).</p>
Full article ">Figure 3
<p>The expected base value <math display="inline"><semantics> <mrow> <mi>E</mi> <mo>[</mo> <mi>f</mi> <mo>(</mo> <mi>z</mi> <mo>)</mo> <mo>]</mo> </mrow> </semantics></math> is the predicted value of the model without any known features and <math display="inline"><semantics> <mrow> <mi>f</mi> <mo>(</mo> <mi>x</mi> <mo>)</mo> </mrow> </semantics></math> is the current output of the model given the input <span class="html-italic">x</span>. The diagram shows how SHAP values attributed to each feature change the expected model prediction when conditioning on that feature.</p>
Full article ">Figure 4
<p>The proposed network architecture. The model serves for both face segmentation and real/fake classification tasks. The model trained this way is used first to infer on the ArtiFact dataset, obtaining its finely cropped version. In a second, separate and successive, phase—represented by the flow ending at the red dotted rectangle labeled ’2’—the model performs the classification task on both versions of the dataset.</p>
Full article ">Figure 5
<p>Application of the segmentation module to some StyleGAN2 generated images. The segmentation process provides finely cropped facial images, isolating the main subjects and blackening the background. The same approach is followed for both the generated and the real samples of the dataset, providing alternative and separate versions of the training samples to ensure that the model is only focusing on faces without the help of the background.</p>
Full article ">Figure 6
<p>Application of SHAP to four test images. For each sample, the input image and the corresponding SHAP values for the coalitions of pixels are reported. Positive SHAP values (red coalitions) indicate that the group of pixels contributes positively toward the model’s prediction, while negative SHAP values (blue coalitions), indicate a negative contribution. In many images, the background plays a crucial role in the decision, both in negative and positive ways (e.g., in the top right and bottom left figures, respectively), thus validating the initial claim and the classification results.</p>
Full article ">
21 pages, 3077 KiB  
Article
AISMSNet: Advanced Image Splicing Manipulation Identification Based on Siamese Networks
by Ana Elena Ramirez-Rodriguez, Rodrigo Eduardo Arevalo-Ancona, Hector Perez-Meana, Manuel Cedillo-Hernandez and Mariko Nakano-Miyatake
Appl. Sci. 2024, 14(13), 5545; https://doi.org/10.3390/app14135545 - 26 Jun 2024
Cited by 1 | Viewed by 929
Abstract
The exponential surge in specialized image editing software has intensified visual forgery, with splicing attacks emerging as a popular forgery technique. In this context, Siamese neural networks are a remarkable tool in pattern identification for detecting image manipulations. This paper introduces a deep [...] Read more.
The exponential surge in specialized image editing software has intensified visual forgery, with splicing attacks emerging as a popular forgery technique. In this context, Siamese neural networks are a remarkable tool in pattern identification for detecting image manipulations. This paper introduces a deep learning approach for splicing detection based on a Siamese neural network tailored to identifying manipulated image regions. The Siamese neural network learns unique features of specific image areas and detects tampered regions through feature comparison. This architecture employs two identical branches with shared weights and image features to compare image blocks and identify tampered areas. Subsequently, a K-means algorithm is applied to identify similar centroids and determine the precise localization of duplicated regions in the image. The experimental results encompass various splicing attacks to assess effectiveness, demonstrating a high accuracy of 98.6% and a precision of 97.5% for splicing manipulation detection. This study presents an advanced splicing image forgery detection and localization algorithm, showcasing its efficacy through comprehensive experiments. Full article
Show Figures

Figure 1

Figure 1
<p>General diagram of the proposed image splicing detection method.</p>
Full article ">Figure 2
<p>Siamese neural network architecture.</p>
Full article ">Figure 3
<p>(<b>a</b>) Original image. (<b>b</b>) Image splicing tampering. (<b>c</b>) Ground truth. (<b>d</b>) Siamese neural network splicing detection. (<b>e</b>) Siamese neural network splicing detection with K-means algorithm refinement.</p>
Full article ">Figure 4
<p>Some examples of original (O) and tampered (T) images contained in the databases used for evaluating the proposed algorithm.</p>
Full article ">Figure 4 Cont.
<p>Some examples of original (O) and tampered (T) images contained in the databases used for evaluating the proposed algorithm.</p>
Full article ">Figure 5
<p>Splicing detection using the proposed systems with different databases.</p>
Full article ">Figure 5 Cont.
<p>Splicing detection using the proposed systems with different databases.</p>
Full article ">Figure 6
<p>Evaluation performance of proposed algorithm: (<b>a</b>) accuracy, (<b>b</b>) precision, (<b>c</b>) <span class="html-italic">recall</span>, (<b>d</b>) <span class="html-italic">F1</span> score, and (<b>e</b>) mean-squared error.</p>
Full article ">Figure 7
<p>Processing time comparison.</p>
Full article ">
21 pages, 3522 KiB  
Article
LBRT: Local-Information-Refined Transformer for Image Copy–Move Forgery Detection
by Peng Liang, Ziyuan Li, Hang Tu and Huimin Zhao
Sensors 2024, 24(13), 4143; https://doi.org/10.3390/s24134143 - 26 Jun 2024
Viewed by 874
Abstract
The current deep learning methods for copy–move forgery detection (CMFD) are mostly based on deep convolutional neural networks, which frequently discard a large amount of detail information throughout convolutional feature extraction and have poor long-range information extraction capabilities. The Transformer structure is adept [...] Read more.
The current deep learning methods for copy–move forgery detection (CMFD) are mostly based on deep convolutional neural networks, which frequently discard a large amount of detail information throughout convolutional feature extraction and have poor long-range information extraction capabilities. The Transformer structure is adept at modeling global context information, but the patch-wise self-attention calculation still neglects the extraction of details in local regions that have been tampered with. A local-information-refined dual-branch network, LBRT (Local Branch Refinement Transformer), is designed in this study. It performs Transformer encoding on the global patches segmented from the image and local patches re-segmented from the global patches using a global modeling branch and a local refinement branch, respectively. The self-attention features from both branches are precisely fused, and the fused feature map is then up-sampled and decoded. Therefore, LBRT considers both global semantic information modeling and local detail information refinement. The experimental results show that LBRT outperforms several state-of-the-art CMFD methods on the USCISI dataset, CASIA CMFD dataset, and DEFACTO CMFD dataset. Full article
(This article belongs to the Special Issue Image Processing and Analysis for Object Detection: 2nd Edition)
Show Figures

Figure 1

Figure 1
<p>Examples of copy–move forgery images.</p>
Full article ">Figure 2
<p>Architecture of the proposed LBRT.</p>
Full article ">Figure 3
<p>Architecture of the global context modeling branch.</p>
Full article ">Figure 4
<p>Architecture of the Transformer encoder block.</p>
Full article ">Figure 5
<p>Architecture of the local refinement branch.</p>
Full article ">Figure 6
<p>Architecture of the feature fusion module and decoder.</p>
Full article ">Figure 7
<p>Visualization examples on the USCISI test set.</p>
Full article ">Figure 8
<p>Visualization examples on the CASIA CMFD test set.</p>
Full article ">Figure 9
<p>Visualization examples on the DEFACTO CMFD test set.</p>
Full article ">Figure 10
<p>Visualization of self-attention heatmaps on three copy–move forgery images.</p>
Full article ">
14 pages, 761 KiB  
Article
Online Signature Biometrics for Mobile Devices
by Katarzyna Roszczewska and Ewa Niewiadomska-Szynkiewicz
Sensors 2024, 24(11), 3524; https://doi.org/10.3390/s24113524 - 30 May 2024
Viewed by 547
Abstract
This paper addresses issues concerning biometric authentication based on handwritten signatures. Our research aimed to check whether a handwritten signature acquired with a mobile device can effectively verify a user’s identity. We present a novel online signature verification method using coordinates of points [...] Read more.
This paper addresses issues concerning biometric authentication based on handwritten signatures. Our research aimed to check whether a handwritten signature acquired with a mobile device can effectively verify a user’s identity. We present a novel online signature verification method using coordinates of points and pressure values at each point collected with a mobile device. Convolutional neural networks are used for signature verification. In this paper, three neural network models are investigated, i.e., two self-made light SigNet and SigNetExt models and the VGG-16 model commonly used in image processing. The convolutional neural networks aim to determine whether the acquired signature sample matches the class declared by the signer. Thus, the scenario of closed set verification is performed. The effectiveness of our method was tested on signatures acquired with mobile phones. We used the subset of the multimodal database, MobiBits, that was captured using a custom-made application and consists of samples acquired from 53 people of diverse ages. The experimental results on accurate data demonstrate that developed architectures of deep neural networks can be successfully used for online handwritten signature verification. We achieved an equal error rate (EER) of 0.63% for random forgeries and 6.66% for skilled forgeries. Full article
(This article belongs to the Special Issue Sensor-Based Behavioral Biometrics)
Show Figures

Figure 1

Figure 1
<p>SigNet network architecture diagram.</p>
Full article ">Figure 2
<p>An example of a genuine signature (<b>top</b>) and its skilled forgery (<b>bottom</b>).</p>
Full article ">Figure 3
<p>Number of samples for each class: genuine signatures (<b>left</b>) and skilled forgeries (<b>right</b>).</p>
Full article ">Figure 4
<p>Handwritten signatures represented as images (pixels correspond to pressure values), scaled to the SigNet input data size.</p>
Full article ">Figure 5
<p>Validation of SigNet, SigNetExt and VGG-16 on the TrainSet dataset. Averaged ROC curves for 20 rounds of cross-validation.</p>
Full article ">Figure 6
<p>Validation of SigNet, SigNetExt and VGG-16 on the ValSet dataset. Averaged ROC curves for 20 rounds of cross-validation.</p>
Full article ">Figure 7
<p>Testing of SigNet, SigNetExt, and VGG-16 on the TestSet dataset. Averaged ROC curves for 20 rounds of cross-validation.</p>
Full article ">
28 pages, 14344 KiB  
Article
Evaluation of Classification Performance of New Layered Convolutional Neural Network Architecture on Offline Handwritten Signature Images
by Yasin Ozkan and Pakize Erdogmus
Symmetry 2024, 16(6), 649; https://doi.org/10.3390/sym16060649 - 23 May 2024
Viewed by 760
Abstract
While there are many verification studies on signature images using deep learning algorithms in the literature, there is a lack of studies on the classification of signature images. Signatures are used as a means of identification for banking, security controls, symmetry, certificates, and [...] Read more.
While there are many verification studies on signature images using deep learning algorithms in the literature, there is a lack of studies on the classification of signature images. Signatures are used as a means of identification for banking, security controls, symmetry, certificates, and contracts. In this study, the aim was to design network architectures that work very fast in areas that require only signature images. For this purpose, a new Si-CNN network architecture with existing layers was designed. Afterwards, a new loss function and layer (Si-CL), a novel architecture using Si-CL as classification layer in Si-CNN to increase the performance of this architecture, was designed. This architecture was called Si-CNN+NC (New Classification). Si-CNN and Si-CNN+NC were trained with two datasets. The first dataset which was used for training is the “C-Signatures” (Classification Signatures) dataset, which was created to test these networks. The second dataset is the “Cedar” dataset, which is a benchmark dataset. The number of classes and sample numbers in the two datasets are symmetrical with each other. To compare the performance of the trained networks, four of the most well-known pre-trained networks, GoogleNet, DenseNet201, Inceptionv3, and ResNet50, were also trained with the two datasets with transfer learning. The findings of the study showed that the proposed network models can learn features from two different handwritten signature images and achieve higher accuracy than other benchmark models. The test success of the trained networks showed that the Si-CNN+NC network outperforms the others, in terms of both accuracy and speed. Finally, Si-CNN and Si-CNN+NC networks were trained with the gold standard dataset MNIST and showed superior performance. Due to its superior performance, Si-CNN and Si-CNN+NC can be used by signature experts as an aid in a variety of applications, including criminal detection and forgery. Full article
(This article belongs to the Special Issue Symmetry/Asymmetry in Neural Networks)
Show Figures

Figure 1

Figure 1
<p>An illustration of the preprocessing phase of the proposed model.</p>
Full article ">Figure 2
<p>An overview of the “C-Signatures” dataset.</p>
Full article ">Figure 3
<p>An overview of the “Cedar Signature” dataset.</p>
Full article ">Figure 4
<p>An overview of the MNIST dataset.</p>
Full article ">Figure 5
<p>Block diagram of the proposed method.</p>
Full article ">Figure 6
<p>Image of the proposed Si-CNN and Si-CNN+NC models.</p>
Full article ">Figure 7
<p>An example of how benchmarking models are built.</p>
Full article ">Figure 8
<p>Accuracy graph of Si-CNN model with “C-Signatures” dataset.</p>
Full article ">Figure 9
<p>Accuracy graph of Si-CNN+NC model with “C-Signatures” dataset.</p>
Full article ">Figure 10
<p>Accuracy plots of benchmark models with “C-Signatures” dataset.</p>
Full article ">Figure 11
<p>Si-CNN model’s confusion matrix for the dataset “C-Signatures”.</p>
Full article ">Figure 12
<p>Si-CNN+NC model’s confusion matrix for the dataset “C-Signatures”.</p>
Full article ">Figure 13
<p>Graph of the benchmark models’ accuracy results on the “C-Signatures” dataset.</p>
Full article ">Figure 14
<p>Accuracy graph of Si-CNN model during training with “Cedar” dataset.</p>
Full article ">Figure 15
<p>Accuracy graph of Si-CNN+NC model during training with “Cedar” dataset.</p>
Full article ">Figure 16
<p>Accuracy plots of benchmarking models with “Cedar” dataset.</p>
Full article ">Figure 17
<p>Si-CNN model’s confusion matrix for the dataset “Cedar”.</p>
Full article ">Figure 18
<p>Si-CNN+NC model’s confusion matrix for the dataset “Cedar”.</p>
Full article ">Figure 19
<p>Graph of the benchmark models’ accuracy results on the “Cedar” dataset.</p>
Full article ">Figure 20
<p>Accuracy graph of Si-CNN model during training with MNIST dataset.</p>
Full article ">Figure 21
<p>Accuracy graph of Si-CNN+NC model during training with MNIST dataset.</p>
Full article ">Figure 22
<p>Si-CNN model’s confusion matrix for the dataset MNIST.</p>
Full article ">Figure 23
<p>Si-CNN+NC model’s confusion matrix for the dataset MNIST.</p>
Full article ">
21 pages, 398 KiB  
Article
A Unique Identification-Oriented Black-Box Watermarking Scheme for Deep Classification Neural Networks
by Mouke Mo, Chuntao Wang and Shan Bian
Symmetry 2024, 16(3), 299; https://doi.org/10.3390/sym16030299 - 4 Mar 2024
Viewed by 1205
Abstract
Given the substantial value and considerable training costs associated with deep neural network models, the field of deep neural network model watermarking has come to the forefront. While black-box model watermarking has made commendable strides, the current methodology for constructing poisoned images in [...] Read more.
Given the substantial value and considerable training costs associated with deep neural network models, the field of deep neural network model watermarking has come to the forefront. While black-box model watermarking has made commendable strides, the current methodology for constructing poisoned images in the existing literature is simplistic and susceptible to forgery. Notably, there is a scarcity of black-box model watermarking techniques capable of discerning a unique user in a multi-user model distribution setting. For this reason, this paper proposes a novel black-box model watermarking method for unique identity identification, which is denoted as the ID watermarking of neural networks (IDwNet). Specifically, to enhance the distinguishability of deep neural network models in multi-user scenarios and mitigate the likelihood of poisoned image counterfeiting, this study develops a discrete cosine transform (DCT) and singular value decomposition (SVD)-based symmetrical embedding method to form the poisoned image. As this ID embedding method leads to indistinguishable deep features, the study constructs a poisoned adversary training strategy by simultaneously inputting clean images, poisoned images with the correct ID, and poisoned adversary images with incorrect IDs to train a deep neural network. Extensive simulation experiments show that the proposed scheme achieves excellent invisibility for the concealed ID, surpassing remarkably the state-of-the-art. In addition, the proposed scheme obtains a validation success rate exceeding 99% for the poisoned images at the cost of a marginal classification accuracy reduction of less than 0.5%. Moreover, even though there is only a 1-bit discrepancy between IDs, the proposed scheme still results in an accurate validation of user copyright. These results indicate that the proposed scheme is promising. Full article
(This article belongs to the Section Computer)
Show Figures

Figure 1

Figure 1
<p>Schematic diagram of the proposed IDwNet. (<b>a</b>) Training stage; (<b>b</b>) Test stage.</p>
Full article ">Figure 2
<p>Effect of toxicity rate on performance.</p>
Full article ">Figure 3
<p>Response performance for different information poisoned images with/without adversary training mode, where adv and non-adv denote the model watermarking response performance with/without adversary training strategy, respectively.</p>
Full article ">Figure 4
<p>Poisoned images and residual plots for GTSRB dataset concerning the original clean image are presented for various black-box model watermarking schemes. Moving from left to right, the top images comprise the original clean image, the poisoned image generated by Patch, Blend, SIG, WaNet, and the newly proposed IDwNet. The bottom row displays a magnified residual map, where the magnitude of the residual pixels is amplified by a factor of 5 compared to the original.</p>
Full article ">Figure 5
<p>Poisoned images and residual plots for CIFAR-10 dataset concerning the original clean image are presented for various black-box model watermarking schemes. Moving from left to right, the top images comprise the original clean image, the poisoned image generated by Patch, Blend, SIG, WaNet, and the newly proposed IDwNet. The bottom row displays a magnified residual map, where the magnitude of the residual pixels is amplified by a factor of 5 compared to the original.</p>
Full article ">
42 pages, 24044 KiB  
Review
Image Inpainting Forgery Detection: A Review
by Adrian-Alin Barglazan, Remus Brad and Constantin Constantinescu
J. Imaging 2024, 10(2), 42; https://doi.org/10.3390/jimaging10020042 - 2 Feb 2024
Cited by 2 | Viewed by 3492
Abstract
In recent years, significant advancements in the field of machine learning have influenced the domain of image restoration. While these technological advancements present prospects for improving the quality of images, they also present difficulties, particularly the proliferation of manipulated or counterfeit multimedia information [...] Read more.
In recent years, significant advancements in the field of machine learning have influenced the domain of image restoration. While these technological advancements present prospects for improving the quality of images, they also present difficulties, particularly the proliferation of manipulated or counterfeit multimedia information on the internet. The objective of this paper is to provide a comprehensive review of existing inpainting algorithms and forgery detections, with a specific emphasis on techniques that are designed for the purpose of removing objects from digital images. In this study, we will examine various techniques encompassing conventional texture synthesis methods as well as those based on neural networks. Furthermore, we will present the artifacts frequently introduced by the inpainting procedure and assess the state-of-the-art technology for detecting such modifications. Lastly, we shall look at the available datasets and how the methods compare with each other. Having covered all the above, the outcome of this study is to provide a comprehensive perspective on the abilities and constraints of detecting object removal via the inpainting procedure in images. Full article
Show Figures

Figure 1

Figure 1
<p>Trends in forgery detection during the last years.</p>
Full article ">Figure 2
<p>Overview of image forgery detection.</p>
Full article ">Figure 3
<p>Original image and the applied mask.</p>
Full article ">Figure 4
<p>Image inpainted using Cahn–Hilliard.</p>
Full article ">Figure 5
<p>Zoomed area to emphasize discontinuities, incorrect color, blurring, and over smoothing.</p>
Full article ">Figure 6
<p>On the left side is the reconstructed image, and on the right side is the difference between reconstructed image and original image.</p>
Full article ">Figure 7
<p>Original image, the applied mask, and the TV result.</p>
Full article ">Figure 8
<p>Staircase artifact seen in zoomed area from the above TV output.</p>
Full article ">Figure 9
<p>Original image, the applied mask, and the Criminisi result.</p>
Full article ">Figure 10
<p>Original image, the applied mask, and the Huang result.</p>
Full article ">Figure 11
<p>Reconstructed image by Criminisi vs. that by Huang.</p>
Full article ">Figure 12
<p>Zoomed area with incorrect color continuation.</p>
Full article ">Figure 13
<p>Zoomed area of the Criminisi result in which jagged areas can be seen.</p>
Full article ">Figure 14
<p>Zoomed area of the Huang result in which blurring is presented.</p>
Full article ">Figure 15
<p>Original image, the applied mask, and Huang result on highly texturized area.</p>
Full article ">Figure 16
<p>Zoomed area in which blurring artifacts are present.</p>
Full article ">Figure 17
<p>Original image, the applied mask, and Lama result with blurring and smoothing effect.</p>
Full article ">Figure 18
<p>Original image, the applied mask, and Lama result with inconsistent textures and coloring.</p>
Full article ">Figure 19
<p>Original image, the applied mask, and Lama result with edge artifacts and structural and geometric errors.</p>
Full article ">Figure 20
<p>Original image, the applied mask, and Lama result with repetitive patterns.</p>
Full article ">Figure 21
<p>General steps for copy-paste forgery detection as presented in P. Korus paper.</p>
Full article ">Figure 22
<p>Original and mask image from dataset.</p>
Full article ">Figure 23
<p>Inpainted results: (<b>a</b>) Criminisi, (<b>b</b>) Gimp, (<b>c</b>) Non-local patch, (<b>d</b>) Lama, (<b>e</b>) Mat.</p>
Full article ">Figure 24
<p>Sample artifact introduced by Gimp.</p>
Full article ">Figure 25
<p>DEBI results on inpainted images using Criminisi (<b>a</b>), Gimp (<b>b</b>), Non-local patch (<b>c</b>), Lama (<b>d</b>), Mat (<b>e</b>).</p>
Full article ">Figure 26
<p>Mantranet results on inpainted images using Criminisi (<b>a</b>), Gimp (<b>b</b>), Non-local patch (<b>c</b>), Lama (<b>d</b>), Mat (<b>e</b>).</p>
Full article ">Figure 27
<p>IID results on inpainted images using Criminisi (<b>a</b>), Gimp (<b>b</b>), Non-local patch (<b>c</b>), Lama (<b>d</b>), Mat (<b>e</b>).</p>
Full article ">Figure 28
<p>Focal results on inpainted images using Criminisi (<b>a</b>), Gimp (<b>b</b>), Non-local patch (<b>c</b>), Lama (<b>d</b>), Mat (<b>e</b>).</p>
Full article ">Figure 29
<p>Evaluation metrics for the results of DEBI detection method applied on the inpainted Open Images Dataset V7 dataset with the following inpainting methods: Criminisi, Gimp, and Non-local Patch.</p>
Full article ">Figure 30
<p>Evaluation metric for the results of CMFD detection method applied on the inpainted Open Images Dataset V7 dataset with the following inpainting methods: Criminisi, Gimp, and Non-local Patch.</p>
Full article ">Figure 31
<p>Evaluation metric for the results of focal detection method applied on the inpainted Open Images Dataset V7 dataset with the following inpainting methods: Criminisi, Gimp, Non-local Patch, Lama, and Mat.</p>
Full article ">Figure 32
<p>Evaluation metric for the results of IID-NET detection method applied on the inpainted Open Images Dataset V7 dataset with the following inpainting methods: Criminisi, Gimp, Non—local patch, Lama, and Mat.</p>
Full article ">Figure 33
<p>Evaluation metric for the results of Mantranet detection method applied on the inpainted Open Images Dataset V7 dataset with the following inpainting methods: Criminisi, Gimp, Non-local Patch, Lama, and Mat.</p>
Full article ">Figure 34
<p>Evaluation metric for the results of PSCC-NET detection method applied on the inpainted Open Images Dataset V7 dataset with the following inpainting methods: Criminisi, Gimp, Non-local patch, Lama, and Mat.</p>
Full article ">Figure 35
<p>Summary of evaluation metrics for the results of IID-NET, focal, Mantranet, and PSCC-NET detection methods applied on the inpainted Open Images Dataset V7 dataset with the following inpainting methods: Criminisi, Gimp, Non-local patch, Lama and Mat.</p>
Full article ">
20 pages, 7937 KiB  
Article
Harmonizing Image Forgery Detection & Localization: Fusion of Complementary Approaches
by Hannes Mareen, Louis De Neve, Peter Lambert and Glenn Van Wallendael
J. Imaging 2024, 10(1), 4; https://doi.org/10.3390/jimaging10010004 - 25 Dec 2023
Cited by 1 | Viewed by 2354
Abstract
Image manipulation is easier than ever, often facilitated using accessible AI-based tools. This poses significant risks when used to disseminate disinformation, false evidence, or fraud, which highlights the need for image forgery detection and localization methods to combat this issue. While some recent [...] Read more.
Image manipulation is easier than ever, often facilitated using accessible AI-based tools. This poses significant risks when used to disseminate disinformation, false evidence, or fraud, which highlights the need for image forgery detection and localization methods to combat this issue. While some recent detection methods demonstrate good performance, there is still a significant gap to be closed to consistently and accurately detect image manipulations in the wild. This paper aims to enhance forgery detection and localization by combining existing detection methods that complement each other. First, we analyze these methods’ complementarity, with an objective measurement of complementariness, and calculation of a target performance value using a theoretical oracle fusion. Then, we propose a novel fusion method that combines the existing methods’ outputs. The proposed fusion method is trained using a Generative Adversarial Network architecture. Our experiments demonstrate improved detection and localization performance on a variety of datasets. Although our fusion method is hindered by a lack of generalization, this is a common problem in supervised learning, and hence a motivation for future work. In conclusion, this work deepens our understanding of forgery detection methods’ complementariness and how to harmonize them. As such, we contribute to better protection against image manipulations and the battle against disinformation. Full article
(This article belongs to the Special Issue Robust Deep Learning Techniques for Multimedia Forensics and Security)
Show Figures

Figure 1

Figure 1
<p>The training procedure of the generator.</p>
Full article ">Figure 2
<p>The training procedure of the discriminator. The dotted box around the discriminator signifies that the weights of these discriminators are shared.</p>
Full article ">Figure 3
<p>Examples of forged images and the corresponding heatmaps by the input forgery localization algorithms and the proposed fusion method.</p>
Full article ">
9 pages, 4934 KiB  
Data Descriptor
Spectrogram Dataset of Korean Smartphone Audio Files Forged Using the “Mix Paste” Command
by Yeongmin Son, Won Jun Kwak and Jae Wan Park
Data 2023, 8(12), 183; https://doi.org/10.3390/data8120183 - 1 Dec 2023
Cited by 2 | Viewed by 1920
Abstract
This study focuses on the field of voice forgery detection, which is increasing in importance owing to the introduction of advanced voice editing technologies and the proliferation of smartphones. This study introduces a unique dataset that was built specifically to identify forgeries created [...] Read more.
This study focuses on the field of voice forgery detection, which is increasing in importance owing to the introduction of advanced voice editing technologies and the proliferation of smartphones. This study introduces a unique dataset that was built specifically to identify forgeries created using the “Mix Paste” technique. This editing technique can overlay audio segments from similar or different environments without creating a new timeframe, making it nearly infeasible to detect forgeries using traditional methods. The dataset consists of 4665 and 45,672 spectrogram images from 1555 original audio files and 15,224 forged audio files, respectively. The original audio was recorded using iPhone and Samsung Galaxy smartphones to ensure a realistic sampling environment. The forged files were created from these recordings and subsequently converted into spectrograms. The dataset also provided the metadata of the original voice files, offering additional context and information that could be used for analysis and detection. This dataset not only fills a gap in existing research but also provides valuable support for developing more efficient deep learning models for voice forgery detection. By addressing the “Mix Paste” technique, the dataset caters to a critical need in voice authentication and forensics, potentially contributing to enhancing security in society. Full article
(This article belongs to the Section Information Systems and Data Management)
Show Figures

Figure 1

Figure 1
<p>Dataset construction process.</p>
Full article ">Figure 2
<p>Proportions of the recording devices used in the study.</p>
Full article ">Figure 3
<p>Distribution of the plain, aspirated, sibilant, and tense consonants in the recording script used in the study.</p>
Full article ">Figure 4
<p>Log spectrogram images: (<b>a</b>) original and (<b>b</b>) forged.</p>
Full article ">Figure 5
<p>Settings for encoding in iTunes.</p>
Full article ">Figure 6
<p>Bona fide bounding box and forged bounding box on linear spectrogram.</p>
Full article ">Figure 7
<p>VGG19-based transfer learning model.</p>
Full article ">Figure 8
<p>Structure of the proposed dataset.</p>
Full article ">
18 pages, 7333 KiB  
Article
Exploring Symmetry in Digital Image Forensics Using a Lightweight Deep-Learning Hybrid Model for Multiple Smoothing Operators
by Saurabh Agarwal and Ki-Hyun Jung
Symmetry 2023, 15(12), 2096; https://doi.org/10.3390/sym15122096 - 22 Nov 2023
Cited by 1 | Viewed by 997
Abstract
Digital images are widely used for informal information sharing, but the rise of fake photos spreading misinformation has raised concerns. To address this challenge, image forensics is employed to verify the authenticity and trustworthiness of these images. In this paper, an efficient scheme [...] Read more.
Digital images are widely used for informal information sharing, but the rise of fake photos spreading misinformation has raised concerns. To address this challenge, image forensics is employed to verify the authenticity and trustworthiness of these images. In this paper, an efficient scheme for detecting commonly used image smoothing operators is presented while maintaining symmetry. A new lightweight deep-learning network is proposed, which is trained with three different optimizers to avoid downsizing to retain critical information. Features are extracted from the activation function of the global average pooling layer in three trained deep networks. These extracted features are then used to train a classification model with an SVM classifier, resulting in significant performance improvements. The proposed scheme is applied to identify averaging, Gaussian, and median filtering with various kernel sizes in small-size images. Experimental analysis is conducted on both uncompressed and JPEG-compressed images, showing superior performance compared to existing methods. Notably, there are substantial improvements in detection accuracy, particularly by 6.50% and 8.20% for 32 × 32 and 64 × 64 images when subjected to JPEG compression at a quality factor of 70. Full article
(This article belongs to the Section Computer)
Show Figures

Figure 1

Figure 1
<p>Non-filtered image and its histogram.</p>
Full article ">Figure 2
<p>Histograms of filtered images.</p>
Full article ">Figure 3
<p>Non-filtered and filtered images with their respective SSIM.</p>
Full article ">Figure 4
<p>SSIM values of filtered images.</p>
Full article ">Figure 5
<p>Probability plot of entropy normal distribution of images.</p>
Full article ">Figure 6
<p>Probability plot of SSIM value normal distribution of images.</p>
Full article ">Figure 7
<p>Framework of proposed Scheme.</p>
Full article ">Figure 8
<p>Layout of proposed deep network.</p>
Full article ">Figure 9
<p>Features extraction process from three trained networks.</p>
Full article ">Figure 10
<p>SVM model training and testing layout.</p>
Full article ">Figure 11
<p>Confusion matrix for image size 64 × 64 using SGDM optimizer.</p>
Full article ">Figure 12
<p>Confusion matrix for image size 64 × 64 using Adam optimizer.</p>
Full article ">Figure 13
<p>Confusion matrix for image size 64 × 64 using RMSprop optimizer.</p>
Full article ">Figure 14
<p>Confusion matrix for image size of 32 × 32 using SGDM optimizer.</p>
Full article ">Figure 15
<p>Confusion matrix for image size 32 × 32 using Adam optimizer.</p>
Full article ">Figure 16
<p>Confusion matrix for image size of 32 × 32 using RMSprop optimizer.</p>
Full article ">Figure 17
<p>Confusion matrix for image size of 64 × 64 with JC = 70 using SGDM optimizer.</p>
Full article ">Figure 18
<p>Confusion matrix for image size of 64 × 64 with JC = 70 using Adam optimizer.</p>
Full article ">Figure 19
<p>Confusion matrix for image size of 64 × 64 with JC = 70 using RMSprop optimizer.</p>
Full article ">Figure 20
<p>MDAC for 32 × 32 size images.</p>
Full article ">Figure 21
<p>MDAC for 64 × 64 size uncompressed images.</p>
Full article ">Figure 22
<p>MDAC for 64 × 64 size compressed images.</p>
Full article ">
15 pages, 7876 KiB  
Article
Detection of AI-Created Images Using Pixel-Wise Feature Extraction and Convolutional Neural Networks
by Fernando Martin-Rodriguez, Rocio Garcia-Mojon and Monica Fernandez-Barciela
Sensors 2023, 23(22), 9037; https://doi.org/10.3390/s23229037 - 8 Nov 2023
Cited by 1 | Viewed by 4508
Abstract
Generative AI has gained enormous interest nowadays due to new applications like ChatGPT, DALL E, Stable Diffusion, and Deepfake. In particular, DALL E, Stable Diffusion, and others (Adobe Firefly, ImagineArt, etc.) can create images from a text prompt and are even able to [...] Read more.
Generative AI has gained enormous interest nowadays due to new applications like ChatGPT, DALL E, Stable Diffusion, and Deepfake. In particular, DALL E, Stable Diffusion, and others (Adobe Firefly, ImagineArt, etc.) can create images from a text prompt and are even able to create photorealistic images. Due to this fact, intense research has been performed to create new image forensics applications able to distinguish between real captured images and videos and artificial ones. Detecting forgeries made with Deepfake is one of the most researched issues. This paper is about another kind of forgery detection. The purpose of this research is to detect photorealistic AI-created images versus real photos coming from a physical camera. Id est, making a binary decision over an image, asking whether it is artificially or naturally created. Artificial images do not need to try to represent any real object, person, or place. For this purpose, techniques that perform a pixel-level feature extraction are used. The first one is Photo Response Non-Uniformity (PRNU). PRNU is a special noise due to imperfections on the camera sensor that is used for source camera identification. The underlying idea is that AI images will have a different PRNU pattern. The second one is error level analysis (ELA). This is another type of feature extraction traditionally used for detecting image editing. ELA is being used nowadays by photographers for the manual detection of AI-created images. Both kinds of features are used to train convolutional neural networks to differentiate between AI images and real photographs. Good results are obtained, achieving accuracy rates of over 95%. Both extraction methods are carefully assessed by computing precision/recall and F1-score measurements. Full article
Show Figures

Figure 1

Figure 1
<p>(<b>a</b>) DALL E 2 image, (<b>b</b>) Stable Diffusion image, and (<b>c</b>) OpenArt image. (<b>d</b>–<b>f</b>) Real photos.</p>
Full article ">Figure 2
<p>(<b>a</b>–<b>c</b>), PRNU patterns computed for AI images of <a href="#sensors-23-09037-f001" class="html-fig">Figure 1</a>. (<b>d</b>–<b>f</b>) are examples of PRNU patterns for real images.</p>
Full article ">Figure 3
<p>(<b>a</b>–<b>c</b>), ELA patterns computed for AI images of <a href="#sensors-23-09037-f001" class="html-fig">Figure 1</a>. (<b>d</b>–<b>f</b>) are examples of ELA patterns for real images.</p>
Full article ">Figure 4
<p>(<b>a</b>) CNN structure used. (<b>b</b>) Layers diagram. (<b>c</b>) Matlab code used to define net structure.</p>
Full article ">Figure 5
<p>CNN training for PRNU patterns. Blue line above: accuracy for training data, black line above: accuracy for validation data. Light brown line below: mean square error for training data, black line below: mean square error for validation data.</p>
Full article ">Figure 6
<p>CNN training for ELA patterns. Blue line above: accuracy for training data, black line above: accuracy for validation data. Light brown line below: mean square error for training data, black line below: mean square error for validation data.</p>
Full article ">Figure 7
<p>CNN training for ELA patterns (extended dataset). Blue line above: accuracy for training data, black line above: accuracy for validation data. Light brown line below: mean square error for training data, black line below: mean square error for validation data.</p>
Full article ">Figure A1
<p>Graphical demo application. (<b>a</b>) Detection of AI image with ELA method. (<b>b</b>) Real photo detection with PRNU method.</p>
Full article ">
12 pages, 8023 KiB  
Article
GP-Net: Image Manipulation Detection and Localization via Long-Range Modeling and Transformers
by Jin Peng, Chengming Liu, Haibo Pang, Xiaomeng Gao, Guozhen Cheng and Bing Hao
Appl. Sci. 2023, 13(21), 12053; https://doi.org/10.3390/app132112053 - 5 Nov 2023
Cited by 3 | Viewed by 1609
Abstract
With the rise of image manipulation techniques, an increasing number of individuals find it easy to manipulate image content. Undoubtedly, this presents a significant challenge to the integrity of multimedia data, thereby fueling the advancement of image forgery detection research. A majority of [...] Read more.
With the rise of image manipulation techniques, an increasing number of individuals find it easy to manipulate image content. Undoubtedly, this presents a significant challenge to the integrity of multimedia data, thereby fueling the advancement of image forgery detection research. A majority of current methods employ convolutional neural networks (CNNs) for image manipulation localization, yielding promising outcomes. Nevertheless, CNN-based approaches possess limitations in establishing explicit long-range relationships. Consequently, addressing the image manipulation localization task necessitates a solution that adeptly builds global context while preserving a robust grasp of low-level details. In this paper, we propose GPNet to address this challenge. GPNet combines Transformer and CNN in parallel which can build global dependency and capture low-level details efficiently. Additionally, we devise an effective fusion module referred to as TcFusion, which proficiently amalgamates feature maps generated by both branches. Thorough extensive experiments conducted on diverse datasets showcase that our network outperforms prevailing state-of-the-art manipulation detection and localization approaches. Full article
(This article belongs to the Special Issue Computer Vision and Pattern Recognition Based on Deep Learning)
Show Figures

Figure 1

Figure 1
<p>Image tampering detection and localization often requires capturing traces left on objects. Hence, it is crucial to leverage object-level consistency for effective forgery detection and localization.</p>
Full article ">Figure 2
<p>Overview of the GP-Net architecture for manipulating detection and localization tasks.</p>
Full article ">Figure 3
<p>The structure of the residual propagation and residual feedback.</p>
Full article ">Figure 4
<p>Visual comparisons of original models. From left to right, we present forged images, the predictions of RRUnet, PSCCNet and ours, GT masks.</p>
Full article ">Figure 5
<p>AUC score (%) of our method with different numbers of training images.</p>
Full article ">
17 pages, 5856 KiB  
Article
Detecting Images in Two-Operator Series Manipulation: A Novel Approach Using Transposed Convolution and Information Fusion
by Saurabh Agarwal, Dae-Jea Cho and Ki-Hyun Jung
Symmetry 2023, 15(10), 1898; https://doi.org/10.3390/sym15101898 - 10 Oct 2023
Cited by 1 | Viewed by 809
Abstract
Digital image forensics is a crucial emerging technique, as image editing tools can modify them easily. Most of the latest methods can determine whether a specific operator has edited an image. These methods are suitable for high-resolution uncompressed images. In practice, more than [...] Read more.
Digital image forensics is a crucial emerging technique, as image editing tools can modify them easily. Most of the latest methods can determine whether a specific operator has edited an image. These methods are suitable for high-resolution uncompressed images. In practice, more than one operator is used to modify image contents repeatedly. In this paper, a reliable scheme using information fusion and deep network networks is presented to recognize manipulation operators and the operator’s series on two operators. A transposed convolutional layer improves the performance of low-resolution JPEG compressed images. In addition, a bottleneck technique is utilized to extend the number of transposed convolutional layers. One average pooling layer is employed to preserve the optimal information flow and evade the overfitting concern among the layers. Moreover, the presented scheme can detect two operator series with various factors without including them in training. The experimental outcomes of the suggested scheme are encouraging and better than the existing schemes due to the availability of sufficient statistical evidence. Full article
(This article belongs to the Section Computer)
Show Figures

Figure 1

Figure 1
<p>UNF and filtered images.</p>
Full article ">Figure 2
<p>Image quality analysis.</p>
Full article ">Figure 3
<p>Effect of operators on images while considering entropy.</p>
Full article ">Figure 4
<p>Framework of the proposed deep network.</p>
Full article ">Figure 5
<p>Bottleneck approach.</p>
Full article ">Figure 6
<p>Information fusion.</p>
Full article ">Figure 7
<p>Training and testing process.</p>
Full article ">Figure 8
<p>Confusion matrix of u = GAB_1.0 and v = USM_3.0.</p>
Full article ">Figure 9
<p>Comparative analysis with [<a href="#B24-symmetry-15-01898" class="html-bibr">24</a>,<a href="#B25-symmetry-15-01898" class="html-bibr">25</a>,<a href="#B26-symmetry-15-01898" class="html-bibr">26</a>] schemes for uncompressed images.</p>
Full article ">Figure 10
<p>Comparative analysis with [<a href="#B24-symmetry-15-01898" class="html-bibr">24</a>,<a href="#B25-symmetry-15-01898" class="html-bibr">25</a>,<a href="#B26-symmetry-15-01898" class="html-bibr">26</a>] schemes of compressed images.</p>
Full article ">
Back to TopTop