Nothing Special   »   [go: up one dir, main page]

You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (384)

Search Parameters:
Keywords = watermark

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
26 pages, 1164 KiB  
Review
Digital Watermarking Technology for AI-Generated Images: A Survey
by Huixin Luo, Li Li and Juncheng Li
Mathematics 2025, 13(4), 651; https://doi.org/10.3390/math13040651 - 16 Feb 2025
Abstract
The rapid advancement of AI-generated content (AIGC) has significantly improved the realism and accessibility of synthetic images. While large image generation models offer immense potential in creative industries, they also introduce serious challenges, including copyright infringement, content authentication, and the traceability of generated [...] Read more.
The rapid advancement of AI-generated content (AIGC) has significantly improved the realism and accessibility of synthetic images. While large image generation models offer immense potential in creative industries, they also introduce serious challenges, including copyright infringement, content authentication, and the traceability of generated images. Digital watermarking has emerged as a promising approach to address these concerns by embedding imperceptible yet detectable signatures into generated images. This survey provides a comprehensive review of three core areas: (1) the evolution of image generation technologies, highlighting key milestones such as the transition from GANs to diffusion models; (2) traditional and state-of-the-art digital image watermarking algorithms, encompassing spatial domain, transform domain, and deep learning-based approaches; (3) watermarking methods specific to AIGC, including ownership authentication of AI model and diffusion model, and watermarking of AI-generated images. Additionally, we examine common performance evaluation metrics used in this field, such as watermark capacity, watermark detection accuracy, fidelity, and robustness. Finally, we discuss the unresolved issues and propose several potential directions for future research. We look forward to this paper offering valuable reference for academics in the field of AIGC watermarking and related fields. Full article
Show Figures

Figure 1

Figure 1
<p>The timeline for technologies of watermarking and image generation [<a href="#B42-mathematics-13-00651" class="html-bibr">42</a>,<a href="#B43-mathematics-13-00651" class="html-bibr">43</a>,<a href="#B44-mathematics-13-00651" class="html-bibr">44</a>,<a href="#B45-mathematics-13-00651" class="html-bibr">45</a>,<a href="#B46-mathematics-13-00651" class="html-bibr">46</a>,<a href="#B47-mathematics-13-00651" class="html-bibr">47</a>,<a href="#B48-mathematics-13-00651" class="html-bibr">48</a>,<a href="#B49-mathematics-13-00651" class="html-bibr">49</a>].</p>
Full article ">Figure 2
<p>The basic process of digital image watermarking scheme.</p>
Full article ">Figure 3
<p>White-box model watermarking and black-box model watermarking.</p>
Full article ">Figure 4
<p>Illustration of different watermark placement with the Stable Diffusion model, including adding to the initial noise and the latent space.</p>
Full article ">Figure 5
<p>The classification of digital watermarking technology for AIGC, including the generated images and the AI models.</p>
Full article ">
26 pages, 17178 KiB  
Article
An Encrypted Speech Integrity Authentication Method: Focus on Fine-Grained Tampering Detection and Tampering Recovery Under High Tamper Ratios
by Fujiu Xu, Jianqiang Li and Xi Xu
Mathematics 2025, 13(4), 573; https://doi.org/10.3390/math13040573 - 9 Feb 2025
Abstract
With the increasing amount of cloud-based speech files, the privacy protection of speech files faces significant challenges. Therefore, integrity authentication of speech files is crucial, and there are two pivotal problems: (1) how to achieve fine-grained and highly accurate tampering detection and (2) [...] Read more.
With the increasing amount of cloud-based speech files, the privacy protection of speech files faces significant challenges. Therefore, integrity authentication of speech files is crucial, and there are two pivotal problems: (1) how to achieve fine-grained and highly accurate tampering detection and (2) how to perform high-quality tampering recovery under high tampering ratios. Tampering detection methods and tampering recovery methods of existing speech integrity authentication are mutually balanced, and most tampering recovery methods are carried out under ideal tampering conditions. This paper proposes an encrypted speech integrity authentication method that can simultaneously address both of problems, and its main contributions are as follows: (1) A 2-least significant bit (2-LSB)-based dual fragile watermarking method is proposed to improve tampering detection performance. This method constructs correlations between encrypted speech sampling points by 2-LSB-based fragile watermarking embedding method and achieves low-error tampering detection of tampered sampling points based on four types of fragile watermarkings. (2) A speech self-recovery model based on residual recovery-based linear interpolation (R2-Lerp) is proposed to achieve tampering recovery under high tampering ratios. This method constructs the model based on the correlation between tampered sampling points and their surrounding sampling points and refines the scenarios of the model according to the tampering situation of the sampling points, with experimental results showing that the recovered speech exhibits improved auditory quality and intelligibility. (3) A scrambling encryption algorithm based on the Lorenz mapping is proposed as the speech encryption method. This method scrambles the speech sampling points several times through 4-dimensional chaotic sequence, with experimental results showing that this method not only ensures security but also slightly improves the effect of tampering recovery. Full article
(This article belongs to the Section E1: Mathematics and Computer Science)
Show Figures

Figure 1

Figure 1
<p>Principles of existing tampering detection methods: TFM extracts the original speech features to construct the database and extracts the authenticated speech features for authentication during tampering detection. TWM constructs the synchronized information as original watermarking and, then, embeds and extracts the watermarking of the authenticated speech for authentication with original watermarking during tampering detection.</p>
Full article ">Figure 2
<p>Principles of existing tampering recovery methods: TRM encodes the original speech as compressed recovery information and embeds and decodes the extracted watermarking in the tampered speech as recovered speech during tampering recovery. TCM recovers the tampered sampling points by correlation between sampling points during tampering recovery.</p>
Full article ">Figure 3
<p>The 2-LSB-based dual fragile watermarking method.</p>
Full article ">Figure 4
<p>Advantages of the proposed speech self-recovery model compared to existing methods [<a href="#B23-mathematics-13-00573" class="html-bibr">23</a>,<a href="#B24-mathematics-13-00573" class="html-bibr">24</a>].</p>
Full article ">Figure 5
<p>The flowchart of the encrypted speech integrity authentication method.</p>
Full article ">Figure 6
<p>Construction of the original tampering detection watermarking.</p>
Full article ">Figure 7
<p>The proposed tampering recovery scenarios.</p>
Full article ">Figure 8
<p>(<b>a</b>) The original speech waveform. (<b>b</b>) The original speech spectrogram. (<b>c</b>) The watermarked speech waveform. (<b>d</b>) The watermarked speech spectrogram. Waveforms and spectrograms of original speech and watermarked speech.</p>
Full article ">Figure 8 Cont.
<p>(<b>a</b>) The original speech waveform. (<b>b</b>) The original speech spectrogram. (<b>c</b>) The watermarked speech waveform. (<b>d</b>) The watermarked speech spectrogram. Waveforms and spectrograms of original speech and watermarked speech.</p>
Full article ">Figure 9
<p>(<b>a</b>) The speech waveform of mute attack. (<b>b</b>) The localization of tampered speech. Waveform and tampering localization after the mute attack of the test speech.</p>
Full article ">Figure 10
<p>(<b>a</b>) The speech waveform of substitution attack. (<b>b</b>) The localization of tampered speech. Waveform and tampering localization after the substitution attack of the test speech.</p>
Full article ">Figure 11
<p>(<b>a</b>) The speech waveform of insertion attack. (<b>b</b>) The localization of tampered speech. Waveform and tampering localization after the insertion attack of the test speech.</p>
Full article ">Figure 12
<p>(<b>a</b>) The speech waveform of deletion attack. (<b>b</b>) The localization of tampered speech. Waveform and tampering localization after the deletion attack of the test speech.</p>
Full article ">Figure 13
<p>(<b>a</b>) 1% tampering; (<b>b</b>) 5% tampering; (<b>c</b>) 10% tampering; (<b>d</b>) 30% tampering. (<b>e</b>) 50% tampering. Recovered Speech waveform and spectrogram at different tampering ratios.</p>
Full article ">Figure 13 Cont.
<p>(<b>a</b>) 1% tampering; (<b>b</b>) 5% tampering; (<b>c</b>) 10% tampering; (<b>d</b>) 30% tampering. (<b>e</b>) 50% tampering. Recovered Speech waveform and spectrogram at different tampering ratios.</p>
Full article ">Figure 14
<p>(<b>a</b>) An encrypted speech waveform. (<b>b</b>) An encrypted speech spectrogram. The waveform and spectrogram of encrypted speech.</p>
Full article ">Figure 15
<p>(<b>a</b>) A speech waveform after the right decryption. (<b>b</b>) A speech spectrogram after the right decryption. (<b>c</b>) A speech waveform after the wrong decryption. (<b>d</b>) A speech spectrogram after the wrong decryption. The waveform and spectrogram of the right initial value and the wrong initial value.</p>
Full article ">
23 pages, 5484 KiB  
Article
Template Watermarking Algorithm for Remote Sensing Images Based on Semantic Segmentation and Ellipse-Fitting
by Xuanyuan Cao, Wei Zhang, Qifei Zhou, Changqing Zhu and Na Ren
Remote Sens. 2025, 17(3), 502; https://doi.org/10.3390/rs17030502 - 31 Jan 2025
Abstract
This study presents a ring template watermarking method utilizing semantic segmentation and elliptical fitting to address the inadequate resilience of digital watermarking techniques for remote sensing images against geometric attacks and affine transformations. The approach employs a convolutional neural network to determine the [...] Read more.
This study presents a ring template watermarking method utilizing semantic segmentation and elliptical fitting to address the inadequate resilience of digital watermarking techniques for remote sensing images against geometric attacks and affine transformations. The approach employs a convolutional neural network to determine the coverage position of the annular template watermark automatically. Subsequently, it applies the least squares approach to align with the relevant elliptic curve of the annular watermark, facilitating the restoration of the watermark template post-deformation due to an attack. Ultimately, it acquires the watermark information by analyzing the binarized image according to the coordinates. The experimental results indicate that, despite various geometric and affine modification attacks, the NC mean value of watermark extraction exceeds 0.83, and the PSNR value surpasses 35, thereby ensuring substantial invisibility and enhanced robustness. In addition, the methods presented in this paper provide useful references for imaging data in other fields. Full article
Show Figures

Figure 1

Figure 1
<p>Main framework of the algorithm.</p>
Full article ">Figure 2
<p>U-Net network structure.</p>
Full article ">Figure 3
<p>Ring template watermark.</p>
Full article ">Figure 4
<p>Spectrogram of ring template watermark.</p>
Full article ">Figure 5
<p>Original spectrogram (<b>a</b>) with mask (<b>b</b>).</p>
Full article ">Figure 6
<p>Loss curve.</p>
Full article ">Figure 7
<p>Model test result (<b>a</b>) with binarized image (<b>b</b>).</p>
Full article ">Figure 8
<p>Ellipse fitting result.</p>
Full article ">Figure 9
<p>Elliptic curve repositioning.</p>
Full article ">Figure 10
<p>Experimental remote sensing imagery data.</p>
Full article ">Figure 11
<p>Method A ring watermark template (<b>a</b>) and Method B watermark information (<b>b</b>).</p>
Full article ">Figure 12
<p>Results of the rotation attacks.</p>
Full article ">Figure 13
<p>Images after scaling attacks.</p>
Full article ">Figure 14
<p>Results of scaling attacks.</p>
Full article ">Figure 15
<p>Image after cropping attacks.</p>
Full article ">Figure 16
<p>Results of cropping attacks.</p>
Full article ">Figure 17
<p>Affine transformation matrices.</p>
Full article ">Figure 18
<p>Results of affine attacks.</p>
Full article ">Figure 19
<p>Effects of watermark embedding strength q and watermark radius R on PSNR.</p>
Full article ">Figure 20
<p>Effects of the number of watermark information bits “1” and watermark radius R on PSNR.</p>
Full article ">Figure 21
<p>Effects of watermark embedding strength q and watermark radius R on NC.</p>
Full article ">Figure 22
<p>Effects of the number of watermark bits “1” and watermark radius R on NC.</p>
Full article ">
9 pages, 6245 KiB  
Article
Multi-Instance Zero-Watermarking Algorithm for Vector Geographic Data
by Qifei Zhou, Lin Yan, Zihao Wang, Na Ren and Changqing Zhu
ISPRS Int. J. Geo-Inf. 2025, 14(2), 54; https://doi.org/10.3390/ijgi14020054 - 30 Jan 2025
Abstract
To address the variability and complexity of attack types, this paper proposes a multi-instance zero-watermarking algorithm that goes beyond the conventional one-to-one watermarking approach. Inspired by the class-instance paradigm in object-oriented programming, this algorithm constructs multiple zero watermarks from a single vector geographic [...] Read more.
To address the variability and complexity of attack types, this paper proposes a multi-instance zero-watermarking algorithm that goes beyond the conventional one-to-one watermarking approach. Inspired by the class-instance paradigm in object-oriented programming, this algorithm constructs multiple zero watermarks from a single vector geographic dataset to enhance resilience against diverse attacks. Normalization is applied to eliminate dimensional and deformation inconsistencies, ensuring robustness against non-uniform scaling attacks. Feature triangle construction and angle selection are further utilized to provide resistance to interpolation and compression attacks. Moreover, angular features confer robustness against translation, uniform scaling, and rotation attacks. Experimental results demonstrate the superior robustness of the proposed algorithm, with normalized correlation values consistently maintaining 1.00 across various attack scenarios. Compared with existing methods, the algorithm exhibits superior comprehensive robustness, effectively safeguarding the copyright of vector geographic data. Full article
(This article belongs to the Topic Recent Advances in Security, Privacy, and Trust)
Show Figures

Figure 1

Figure 1
<p>Main framework of the proposed method.</p>
Full article ">Figure 2
<p>Demonstration of feature triangle construction: (<b>a</b>) a polyline with nine vertices and (<b>b</b>) the feature triangle ABC.</p>
Full article ">Figure 3
<p>Waterways dataset used in the experiments.</p>
Full article ">Figure 4
<p>Visualization of attack effects: (<b>a</b>) rotation (60°); (<b>b</b>) rotation (120°); (<b>c</b>) uniform scaling (Sx = Sy = 0.4); (<b>d</b>) non-uniform scaling (Sx = 0.4, Sy = 1); (<b>e</b>) non-uniform scaling (Sx = 0.8, Sy = 1); (<b>f</b>) translation (60 m); (<b>g</b>) interpolation (tolerance = 10 m); and (<b>h</b>) simplification (tolerance = 60 m).</p>
Full article ">Figure 4 Cont.
<p>Visualization of attack effects: (<b>a</b>) rotation (60°); (<b>b</b>) rotation (120°); (<b>c</b>) uniform scaling (Sx = Sy = 0.4); (<b>d</b>) non-uniform scaling (Sx = 0.4, Sy = 1); (<b>e</b>) non-uniform scaling (Sx = 0.8, Sy = 1); (<b>f</b>) translation (60 m); (<b>g</b>) interpolation (tolerance = 10 m); and (<b>h</b>) simplification (tolerance = 60 m).</p>
Full article ">Figure 5
<p>Results of rotation attacks.</p>
Full article ">Figure 6
<p>Results of uniform scaling attacks.</p>
Full article ">Figure 7
<p>Results of non-uniform scaling attacks.</p>
Full article ">Figure 8
<p>Results of translation attacks.</p>
Full article ">Figure 9
<p>Results of interpolation attacks.</p>
Full article ">Figure 10
<p>Results of simplification attacks.</p>
Full article ">Figure 11
<p>Results for individual watermarks under rotation attacks.</p>
Full article ">Figure 12
<p>Results for individual watermarks under non-uniform scaling attacks.</p>
Full article ">Figure 13
<p>Results for individual watermarks under simplification attacks.</p>
Full article ">
22 pages, 10634 KiB  
Article
Copyright Verification and Traceability for Remote Sensing Object Detection Models via Dual Model Watermarking
by Weitong Chen, Xin Xu, Na Ren, Changqing Zhu and Jie Cai
Remote Sens. 2025, 17(3), 481; https://doi.org/10.3390/rs17030481 - 30 Jan 2025
Abstract
Deep learning-based remote sensing object detection (RSOD) models have been widely deployed and commercialized. The commercialization of RSOD models requires the ability to protect their intellectual property (IP) across different platforms and sales channels. However, RSOD models currently face threats related to illegal [...] Read more.
Deep learning-based remote sensing object detection (RSOD) models have been widely deployed and commercialized. The commercialization of RSOD models requires the ability to protect their intellectual property (IP) across different platforms and sales channels. However, RSOD models currently face threats related to illegal copying on untrusted platforms or resale by dishonest buyers. To address this issue, we propose a dual-model watermarking scheme for the copyright verification and leakage tracing of RSOD models. First, we construct trigger samples using an object generation watermark trigger and train them alongside clean samples to implement black-box watermarking. Then, fingerprint information is embedded into a small subset of the model’s critical weights, using a fine-tuning and loss-guided approach. At the copyright verification stage, the presence of a black-box watermark can be confirmed through using the suspect model’s API to make predictions on the trigger samples, thereby determining whether the model is infringing. Once infringement is confirmed, fingerprint information can be further extracted from the model weights to identify the leakage source. Experimental results demonstrate that the proposed method can effectively achieve the copyright verification and traceability of RSOD models without affecting the performance of primary tasks. The watermark shows good robustness against fine-tuning and pruning attacks. Full article
Show Figures

Figure 1

Figure 1
<p>Threats to IP of owners of remote sensing deep learning models.</p>
Full article ">Figure 2
<p>Overview of the proposed copyright verification and traceability scheme for RSOD models. The proposed scheme consists of three main components: (1) embedding black-box model watermarks for copyright verification; (2) embedding identification information for traceability; and (3) performing infringement detection and leakage tracing to verify ownership and hold traitors accountable.</p>
Full article ">Figure 3
<p>Samples embedded with different types of object generation watermark triggers. The three rows demonstrate samples embedded with triggers of White_and_Black_Grids, Red_Airplane, and RS_Logo, respectively. The red box highlights the exact location where the triggers are embedded in the samples.</p>
Full article ">Figure 4
<p>Fingerprint embedding process.</p>
Full article ">Figure 5
<p>Inference results of pirated and non-pirated models. (<b>a</b>) Pirated model. (<b>b</b>) Non-pirated model.</p>
Full article ">Figure 6
<p>The variation in <math display="inline"><semantics> <mi>WSR</mi> </semantics></math> under different training epochs. (<b>a</b>) <math display="inline"><semantics> <mi>WSR</mi> </semantics></math> variation on Faster R-CNN model. (<b>b</b>) <math display="inline"><semantics> <mi>WSR</mi> </semantics></math> variation on YOLOv5 model.</p>
Full article ">Figure 7
<p>Examples of successful watermark validation results for trigger samples.</p>
Full article ">Figure 8
<p>The variation in <math display="inline"><semantics> <mi mathvariant="script">BCR</mi> </semantics></math> during different fine-tuning epochs. (<b>a</b>) <math display="inline"><semantics> <mi mathvariant="script">BCR</mi> </semantics></math> variation on Faster R-CNN model. (<b>b</b>) <math display="inline"><semantics> <mi mathvariant="script">BCR</mi> </semantics></math> variation on YOLOv5 model.</p>
Full article ">Figure 9
<p>Robustness against fine-tuning attacks. (<b>a</b>) Black-box watermark in Faster-RCNN model. (<b>b</b>) Black-box watermark in YOLOv5 model. (<b>c</b>) Fingerprint in Faster-RCNN model. (<b>d</b>) Fingerprint in YOLOv5 model.</p>
Full article ">Figure 10
<p>Robustness against pruning attacks. (<b>a</b>) Faster-RCNN model with NWPU VHR-10 dataset. (<b>b</b>) Faster-RCNN model with RSOD dataset. (<b>c</b>) Faster-RCNN model with LEVIR dataset. (<b>d</b>) YOLOv5 model with NWPU VHR-10 dataset. (<b>e</b>) YOLOv5 model with RSOD dataset. (<b>f</b>) YOLOv5 model with LEVIR dataset.</p>
Full article ">Figure 11
<p>Influence of different configurations on model performance. (<b>a</b>) Trigger sizes. (<b>b</b>) Watermarking rate. (<b>c</b>) Trigger patterns.</p>
Full article ">
30 pages, 5698 KiB  
Article
A Blockchain Copyright Protection Model Based on Vector Map Unique Identification
by Heyan Wang, Nannan Tang, Changqing Zhu, Na Ren and Changhong Wang
ISPRS Int. J. Geo-Inf. 2025, 14(2), 53; https://doi.org/10.3390/ijgi14020053 - 30 Jan 2025
Abstract
Combining blockchain technology with digital watermarking presents an efficient solution for safeguarding vector map files. However, the large data volume and stringent confidentiality requirements of vector maps pose significant challenges for direct registration on blockchain platforms. To overcome these limitations, this paper proposes [...] Read more.
Combining blockchain technology with digital watermarking presents an efficient solution for safeguarding vector map files. However, the large data volume and stringent confidentiality requirements of vector maps pose significant challenges for direct registration on blockchain platforms. To overcome these limitations, this paper proposes a blockchain-based copyright protection model utilizing unique identifiers (BCPM-UI). The model employs a distance ratio-based quantization watermarking algorithm to embed watermark information into vector maps and then generates unique identifiers based on their topological and geometric parameters. These identifiers, rather than the vector maps themselves, are securely registered on the blockchain. To ensure reliable copyright verification, a bit error rate (BER)-based matching algorithm is introduced, enabling accurate comparison between the unique identifiers of suspected infringing data and those stored on the blockchain. Experimental results validate the model’s effectiveness, demonstrating the high uniqueness and robustness of the identifiers generated. Additionally, the proposed approach reduces blockchain storage requirements for map data by a factor of 200, thereby meeting confidentiality standards while maintaining practical applicability in terms of copyright protection for vector maps. Full article
(This article belongs to the Topic Recent Advances in Security, Privacy, and Trust)
Show Figures

Figure 1

Figure 1
<p>Copyright registration process in the BCPM-UI model: feature identifier construction, watermark embedding, and blockchain registration.</p>
Full article ">Figure 2
<p>Antchain combined with IPFS for vector map copyright protection.</p>
Full article ">Figure 3
<p>Schematic diagram of angle feature parameter acquisition.</p>
Full article ">Figure 4
<p>Schematic of nearest neighbor non-intersecting heterogeneous feature query based on Hausdorff distance.</p>
Full article ">Figure 5
<p>Calculation diagram of distance ratio.</p>
Full article ">Figure 6
<p>Quantization mechanism watermark embedding diagram.</p>
Full article ">Figure 7
<p>The dataset used in the experiment: (<b>a</b>) Shanghai dataset; (<b>b</b>) Beijing dataset; (<b>c</b>) Chengdu dataset; (<b>d</b>) Jiangsu dataset; (<b>e</b>) Hangzhou dataset; (<b>f</b>) Nanjing dataset.</p>
Full article ">Figure 8
<p>Watermark images used in the experiment.</p>
Full article ">Figure 9
<p>The amount of memory space occupied in the blockchain.</p>
Full article ">Figure 10
<p>Watermarked vector map: (<b>a</b>) Shanghai dataset; (<b>b</b>) Beijing dataset; (<b>c</b>) Chengdu dataset; (<b>d</b>) Jiangsu dataset; (<b>e</b>) Hangzhou dataset; (<b>f</b>) Nanjing dataset.</p>
Full article ">Figure 11
<p>Robustness experiment results: (<b>a</b>) rotation attack; (<b>b</b>) scaling attack; (<b>c</b>) translation attack; (<b>d</b>) object-add attack; (<b>e</b>) object-delete attack; (<b>f</b>) layer-add attack; (<b>g</b>) layer-delete attack; (<b>h</b>) cropping attack; (<b>i</b>) merge attack.</p>
Full article ">Figure 11 Cont.
<p>Robustness experiment results: (<b>a</b>) rotation attack; (<b>b</b>) scaling attack; (<b>c</b>) translation attack; (<b>d</b>) object-add attack; (<b>e</b>) object-delete attack; (<b>f</b>) layer-add attack; (<b>g</b>) layer-delete attack; (<b>h</b>) cropping attack; (<b>i</b>) merge attack.</p>
Full article ">Figure 12
<p>Experimental results of uniqueness between datasets with different unique identification lengths: (<b>a</b>) Shanghai dataset; (<b>b</b>) Beijing dataset; (<b>c</b>) Chengdu dataset; (<b>d</b>) Jiangsu dataset; (<b>e</b>) Hangzhou dataset; (<b>f</b>) Nanjing dataset.</p>
Full article ">Figure 13
<p>Experimental results of robustness with different unique identification length: (<b>a</b>) cropping attack; (<b>b</b>) layer-delete attack; (<b>c</b>) object-delete attack.</p>
Full article ">Figure 14
<p>Vector map dataset with small data volume: (<b>a</b>) Chongqing dataset; (<b>b</b>) Xi’an dataset.</p>
Full article ">Figure 15
<p>Experimental results of robustness under vector maps with small data volume: (<b>a</b>) object-delete attack; (<b>b</b>) cropping attack.</p>
Full article ">
23 pages, 3403 KiB  
Article
Class-Hidden Client-Side Watermarking in Federated Learning
by Weitong Chen, Chi Zhang, Wei Zhang and Jie Cai
Entropy 2025, 27(2), 134; https://doi.org/10.3390/e27020134 - 27 Jan 2025
Abstract
Federated learning consists of a central aggregator and multiple clients, forming a distributed structure that effectively protects data privacy. However, since all participants can access the global model, the risk of model leakage increases, especially when unreliable participants are involved. To safeguard model [...] Read more.
Federated learning consists of a central aggregator and multiple clients, forming a distributed structure that effectively protects data privacy. However, since all participants can access the global model, the risk of model leakage increases, especially when unreliable participants are involved. To safeguard model copyright while enhancing the robustness and secrecy of the watermark, this paper proposes a client-side watermarking scheme. Specifically, the proposed method introduces an additional watermark class, expanding the output layer of the client model into an N+1-class classifier. The client’s local model is then trained using both the watermark dataset and the local dataset. Notably, before uploading to the server, the parameters of the watermark class are removed from the output layer and stored locally. Additionally, the client uploads amplified parameters to address the potential weakening of the watermark during the aggregation. After aggregation, the global model is distributed to the clients for local training. Through multiple rounds of iteration, the saved watermark parameters are continuously updated until the global model converges. On the MNIST, CIFAR-100, and CIFAR-10 datasets, the watermark detection rates on VGG-16 and ResNet-18 reached 100%. Furthermore, extensive experiments demonstrate that this method has minimal impact on model performance and exhibits strong robustness against pruning and fine-tuning attacks. Full article
(This article belongs to the Special Issue Applications of Information Theory to Machine Learning)
Show Figures

Figure 1

Figure 1
<p>Samples in the watermark dataset.</p>
Full article ">Figure 2
<p>Watermark embedding process.</p>
Full article ">Figure 3
<p>Watermark verification process.</p>
Full article ">Figure 4
<p>The impact of watermark dataset size: (<b>a</b>) The impact on the accuracy of the original task. (<b>b</b>) The impact on watermark detection rate.</p>
Full article ">Figure 5
<p>Comparison of original task accuracy.</p>
Full article ">Figure 6
<p>Accuracy of original task and watermark across different training epochs.</p>
Full article ">Figure 7
<p>Comparison of watermark accuracy across fine-tuning epochs.</p>
Full article ">Figure 8
<p>Comparison of original task accuracy against forging attack.</p>
Full article ">Figure 9
<p>The activation maps and weight distributions of the watermarked model and the clean model on VGG-16.</p>
Full article ">Figure 10
<p>Distribution of prediction results for watermark datasets on clean and watermarked models: (<b>a</b>) Clean model. (<b>b</b>) Unmodified watermarked model. (<b>c</b>) Modified watermarked model.</p>
Full article ">Figure 11
<p>Accuracy of original task and watermark under Non-IID conditions.</p>
Full article ">Figure 12
<p>Accuracy of original task and watermark against fine-tuning under Non-IID conditions.</p>
Full article ">Figure 13
<p>Accuracy of original task and watermark against pruning under Non-IID conditions.</p>
Full article ">Figure 14
<p>Watermark accuracy against overwriting attack under Non-IID conditions.</p>
Full article ">
19 pages, 8990 KiB  
Article
Optimizing Image Watermarking with Dual-Tree Complex Wavelet Transform and Particle Swarm Intelligence for Secure and High-Quality Protection
by Abed Al Raoof Bsoul and Alaa Bani Ismail
Appl. Sci. 2025, 15(3), 1315; https://doi.org/10.3390/app15031315 - 27 Jan 2025
Abstract
Watermarking is a technique used to address issues related to the widespread use of the internet, such as copyright protection, tamper localization, and authentication. However, most watermarking approaches negatively affect the quality of the original image. In this research, we propose an optimized [...] Read more.
Watermarking is a technique used to address issues related to the widespread use of the internet, such as copyright protection, tamper localization, and authentication. However, most watermarking approaches negatively affect the quality of the original image. In this research, we propose an optimized image watermarking approach that utilizes the dual-tree complex wavelet transform and particle swarm optimization algorithm. Our approach focuses on maintaining the highest possible quality of the watermarked image by minimizing any noticeable changes. During the embedding phase, we break down the original image using a technique called dual-tree complex wavelet transform (DTCWT) and then use particle swarm optimization (PSO) to choose specific coefficients. We embed the bits of a binary logo into the least significant bits of these selected coefficients, creating the watermarked image. To extract the watermark, we reverse the embedding process by first decomposing both versions of the input image using DTCWT and extracting the same coefficients to retrieve those corresponding bits (watermark). In our experiments, we used a common dataset from watermarking research to demonstrate the functionality against various watermarked copies and peak signal-to-noise ratio (PSNR) and normalized cross-correlation (NCC) metrics. The PSNR is a measure of how well the watermarked image maintains its original quality, and the NCC reflects how accurately the watermark can be extracted. Our method gives mean PSNR and NCC of 80.50% and 92.51%, respectively. Full article
(This article belongs to the Special Issue Digital Image Processing: Technologies and Applications)
Show Figures

Figure 1

Figure 1
<p>The architecture of a digital image watermarking system.</p>
Full article ">Figure 2
<p>The overall design of the proposed method.</p>
Full article ">Figure 3
<p>The logo used for embedding in this research. The actual logo size is 100 × 112, and the logo is black and white.</p>
Full article ">Figure 4
<p>DTCWT sub-bands using multi-level decomposition.</p>
Full article ">Figure 5
<p>Impact of inertia weight (<span class="html-italic">ω</span>) on computational cost and PSNR.</p>
Full article ">Figure 6
<p>Impact of cognitive component (<span class="html-italic">C</span>1) on computational cost and PSNR.</p>
Full article ">Figure 7
<p>Impact of social component (<span class="html-italic">C</span>2) on computational cost and PSNR.</p>
Full article ">Figure 8
<p>Impact of swarm size on computational cost and PSNR.</p>
Full article ">Figure 9
<p>Impact of number of iterations on computational cost and PSNR.</p>
Full article ">
19 pages, 2333 KiB  
Review
Detection of Manipulations in Digital Images: A Review of Passive and Active Methods Utilizing Deep Learning
by Paweł Duszejko, Tomasz Walczyna and Zbigniew Piotrowski
Appl. Sci. 2025, 15(2), 881; https://doi.org/10.3390/app15020881 - 17 Jan 2025
Viewed by 460
Abstract
The modern society generates vast amounts of digital content, whose credibility plays a pivotal role in shaping public opinion and decision-making processes. The rapid development of social networks and generative technologies, such as deepfakes, significantly increases the risk of disinformation through image manipulation. [...] Read more.
The modern society generates vast amounts of digital content, whose credibility plays a pivotal role in shaping public opinion and decision-making processes. The rapid development of social networks and generative technologies, such as deepfakes, significantly increases the risk of disinformation through image manipulation. This article aims to review methods for verifying images’ integrity, particularly through deep learning techniques, addressing both passive and active approaches. Their effectiveness in various scenarios has been analyzed, highlighting their advantages and limitations. This study reviews the scientific literature and research findings, focusing on techniques that detect image manipulations and localize areas of tampering, utilizing both statistical properties of images and embedded hidden watermarks. Passive methods, based on analyzing the image itself, are versatile and can be applied across a broad range of cases; however, their effectiveness depends on the complexity of the modifications and the characteristics of the image. Active methods, which involve embedding additional information into the image, offer precise detection and localization of changes but require complete control over creating and distributing visual materials. Both approaches have their applications depending on the context and available resources. In the future, a key challenge remains the development of methods resistant to advanced manipulations generated by diffusion models and further leveraging innovations in deep learning to protect the integrity of visual content. Full article
(This article belongs to the Special Issue Integration of AI in Signal and Image Processing)
Show Figures

Figure 1

Figure 1
<p>An example of reality falsification by a totalitarian regime: the removal of Nikolai Yezhov from a photograph with Stalin [<a href="#B4-applsci-15-00881" class="html-bibr">4</a>].</p>
Full article ">Figure 2
<p>ManTra-Net high level architecture.</p>
Full article ">Figure 3
<p>SPAN high level architecture.</p>
Full article ">Figure 4
<p>Asnani et al. [<a href="#B35-applsci-15-00881" class="html-bibr">35</a>] high level architecture.</p>
Full article ">Figure 5
<p>ObjectFormer high level architecture.</p>
Full article ">Figure 6
<p>Zhao et al. [<a href="#B36-applsci-15-00881" class="html-bibr">36</a>] high level architecture.</p>
Full article ">
14 pages, 608 KiB  
Article
TIBW: Task-Independent Backdoor Watermarking with Fine-Tuning Resilience for Pre-Trained Language Models
by Weichuan Mo, Kongyang Chen and Yatie Xiao
Mathematics 2025, 13(2), 272; https://doi.org/10.3390/math13020272 - 15 Jan 2025
Viewed by 527
Abstract
Pre-trained language models such as BERT, GPT-3, and T5 have made significant advancements in natural language processing (NLP). However, their widespread adoption raises concerns about intellectual property (IP) protection, as unauthorized use can undermine innovation. Watermarking has emerged as a promising solution for [...] Read more.
Pre-trained language models such as BERT, GPT-3, and T5 have made significant advancements in natural language processing (NLP). However, their widespread adoption raises concerns about intellectual property (IP) protection, as unauthorized use can undermine innovation. Watermarking has emerged as a promising solution for model ownership verification, but its application to NLP models presents unique challenges, particularly in ensuring robustness against fine-tuning and preventing interference with downstream tasks. This paper presents a novel watermarking scheme, TIBW (Task-Independent Backdoor Watermarking), that embeds robust, task-independent backdoor watermarks into pre-trained language models. By implementing a Trigger–Target Word Pair Search Algorithm that selects trigger–target word pairs with maximal semantic dissimilarity, our approach ensures that the watermark remains effective even after extensive fine-tuning. Additionally, we introduce Parameter Relationship Embedding (PRE) to subtly modify the model’s embedding layer, reinforcing the association between trigger and target words without degrading the model performance. We also design a comprehensive watermark verification process that evaluates task behavior consistency, quantified by the Watermark Embedding Success Rate (WESR). Our experiments across five benchmark NLP tasks demonstrate that the proposed watermarking method maintains a near-baseline performance on clean inputs while achieving a high WESR, outperforming existing baselines in both robustness and stealthiness. Furthermore, the watermark persists reliably even after additional fine-tuning, highlighting its resilience against potential watermark removal attempts. This work provides a secure and reliable IP protection mechanism for NLP models, ensuring watermark integrity across diverse applications. Full article
Show Figures

Figure 1

Figure 1
<p>Process of embedding a watermark into the embedding layer of pre-trained language models and verifying ownership through trigger–target word pairs.</p>
Full article ">Figure 2
<p>t-SNE visualization of trigger and target word embeddings. Points in the lower-right corner represent low-frequency trigger words, while those in the upper-left corner represent high-frequency target words. The clear separation of these two types of words in the embedding space validates the algorithm’s success in maximizing semantic distance.</p>
Full article ">Figure 3
<p>Resilience of the watermark under extended fine-tuning across 5 epochs. The plots show the accuracy/F1-score on clean inputs and WESR on trigger inputs for SST2, Ling-Spam, and SQuAD tasks.</p>
Full article ">
23 pages, 3923 KiB  
Article
A Robust Semi-Blind Watermarking Technology for Resisting JPEG Compression Based on Deep Convolutional Generative Adversarial Networks
by Chin-Feng Lee, Zih-Cyuan Chao, Jau-Ji Shen and Anis Ur Rehman
Symmetry 2025, 17(1), 98; https://doi.org/10.3390/sym17010098 - 10 Jan 2025
Viewed by 458
Abstract
In recent years, the internet has developed rapidly. With the popularity of social media, uploading and backing up digital images has become the norm. A huge number of digital images are circulating on the internet daily, and issues related to information security follow. [...] Read more.
In recent years, the internet has developed rapidly. With the popularity of social media, uploading and backing up digital images has become the norm. A huge number of digital images are circulating on the internet daily, and issues related to information security follow. To protect intellectual property rights, digital watermarking is an indispensable technology. However, the common lossy compression technology in the network transmission process is a big problem for watermarking. This paper describes an innovative semi-blind watermarking method with the use of deep convolutional generative adversarial networks (DCGANs) for hiding and extracting watermarks from JPEG-compressed images. The proposed method achieves an average peak signal-to-noise ratio (PSNR) of 49.99 dB, a structural similarity index (SSIM) of 0.95, and a bit error rate (BER) of 0.008 across varying JPEG quality factors. The process is based on an embedder, decoder, generator, and discriminator. It allows watermarking, decoding, or reconstruction to be symmetric such that there is less distortion and durability is improved. It constructs a specific generator for each image and watermark that is supposed to be protected. Experimental results show that, with the variety of JPEG quality factors, the restored watermark achieves a remarkably low corrupted rate, outstripping recent deep learning-based watermarking methods. Full article
(This article belongs to the Section Computer)
Show Figures

Figure 1

Figure 1
<p>The framework diagram of the proposed model.</p>
Full article ">Figure 2
<p>Part of the image of the USC-SIPI image dataset. (<b>a</b>) Lena, (<b>b</b>) baboon, (<b>c</b>) horse, (<b>d</b>) F16, (<b>e</b>) peppers, (<b>f</b>) Sailboat.</p>
Full article ">Figure 3
<p>Part of the image of the Microsoft COCO dataset. (<b>a</b>) Zebra, (<b>b</b>) boat, (<b>c</b>) plane, (<b>d</b>) cat, (<b>e</b>) motorcycle, (<b>f</b>) old man.</p>
Full article ">Figure 4
<p>Part of the binary image of National Chung Hsing University LOGO and CIFAR. (<b>a</b>) NCHU logo, (<b>b</b>) Man, (<b>c</b>) paw, (<b>d</b>) bike, (<b>e</b>) camel, (<b>f</b>) women.</p>
Full article ">Figure 5
<p>Example of embedder.</p>
Full article ">Figure 6
<p>Example of decoder.</p>
Full article ">Figure 7
<p>Generator architecture diagram.</p>
Full article ">Figure 8
<p>Discriminator architecture diagram.</p>
Full article ">Figure 9
<p>Visualized results of restored (retrieved) watermarks in different situations.</p>
Full article ">Figure 10
<p>Visualization results of restored (retrieved) watermark with different models.</p>
Full article ">Figure 11
<p>Generator loss with different models. (<b>a</b>) Generator loss of the original model. (<b>b</b>) Generator loss of the model DSC-1. (<b>c</b>) Generator loss of the model DSC-2.</p>
Full article ">
24 pages, 886 KiB  
Article
Double Security Level Protection Based on Chaotic Maps and SVD for Medical Images
by Conghuan Ye, Shenglong Tan, Jun Wang, Li Shi, Qiankun Zuo and Bing Xiong
Mathematics 2025, 13(2), 182; https://doi.org/10.3390/math13020182 - 8 Jan 2025
Viewed by 589
Abstract
The widespread distribution of medical images in smart healthcare systems will cause privacy concerns. The unauthorized sharing of decrypted medical images remains uncontrollable, though image encryption can discourage privacy disclosure. This research proposes a double-level security scheme for medical images to overcome this [...] Read more.
The widespread distribution of medical images in smart healthcare systems will cause privacy concerns. The unauthorized sharing of decrypted medical images remains uncontrollable, though image encryption can discourage privacy disclosure. This research proposes a double-level security scheme for medical images to overcome this problem. The proposed joint encryption and watermarking scheme is based on singular-value decomposition (SVD) and chaotic maps. First, three different random sequences are used to encrypt the LL subband in the discrete wavelet transform (DWT) domain; then, HL and LH sub-bands are embedded with watermark information; in the end, we obtain the watermarked and encrypted image with the inverse DWT (IDWT) transform. In this study, SVD is used for watermarking and encryption in the DWT domain. The main originality is that decryption and watermark extraction can be performed separately. Experimental results demonstrate the superiority of the proposed method in key spaces (10225), PSNR (76.2543), and UACI (0.3329). In this implementation, the following key achievements are attained. First, our scheme can meet requests of different security levels. Second, encryption and watermarking can be performed separately. Third, the watermark can be detected in the encrypted domain. Thus, experiment results and security analysis demonstrate the effectiveness of the proposed scheme. Full article
(This article belongs to the Special Issue Mathematical Models in Information Security and Cryptography)
Show Figures

Figure 1

Figure 1
<p>The proposed security scheme.</p>
Full article ">Figure 2
<p>The proposed JEW scheme.</p>
Full article ">Figure 3
<p>Encryption results: (<b>a</b>) original images, (<b>b</b>) encrypted images, (<b>c</b>) decrypted and watermarked images, (<b>d</b>) original histograms, (<b>e</b>) encrypted histograms.</p>
Full article ">Figure 4
<p>Discussion of the selective encryption: (<b>a</b>) original medical images, (<b>b</b>) permutation in the LL subband, (<b>c</b>) diffusion with SVD operation, (<b>d</b>) diffusion with bitxor operation, (<b>e</b>) decrypted images.</p>
Full article ">
16 pages, 3045 KiB  
Article
Reversible Spectral Speech Watermarking with Variable Embedding Locations Against Spectrum-Based Attacks
by Xuping Huang and Akinori Ito
Appl. Sci. 2025, 15(1), 381; https://doi.org/10.3390/app15010381 - 3 Jan 2025
Viewed by 451
Abstract
To guarantee the reliability and integrity of audio, data have been focused on as an essential topic as the fast development of generative AI. Significant progress in machine learning and speech synthesis has increased the potential for audio tampering. In this paper, we [...] Read more.
To guarantee the reliability and integrity of audio, data have been focused on as an essential topic as the fast development of generative AI. Significant progress in machine learning and speech synthesis has increased the potential for audio tampering. In this paper, we focus on the digital watermarking method as a promising method to safeguard the authenticity of audio evidence. Due to the integrity of the original data with probative importance, the algorithm requires reversibility, imperceptibility, and reliability. To meet the requirements, we propose a reversible digital watermarking approach that embeds feature data concentrating in high-frequency intDCT coefficients after transforming data from the time domain into the frequency domain. We explored the appropriate hiding locations against spectrum-based attacks with novel proposed methodologies for spectral expansion for embedding. However, the drawback of fixed expansion is that the stego signal is prone to being detected by a spectral analysis. Therefore, this paper proposes two other new expansion methodologies that embed the data into variable locations—random expansion and adaptive expansion with distortion estimation for embedding—which effectively conceal the watermark’s presence while maintaining high perceptual quality with an average segSNR better than 21.363 dB and average MOS value better than 4.085. Our experimental results demonstrate the efficacy of our proposed method in both sound quality preservation and log-likelihood value, indicating the absolute discontinuity of the spectrogram after embedding is proposed to evaluate the effectiveness of the proposed reversible spectral expansion watermarking algorithm. The result of EER indicated that the adaptive hiding performed best against attacks by spectral analysis. Full article
(This article belongs to the Section Computing and Artificial Intelligence)
Show Figures

Figure 1

Figure 1
<p>Spectrogram difference between original data and stego data, with a red mark on the borderline caused by the watermarking: Ja_m5.wav (<math display="inline"><semantics> <mrow> <mi>N</mi> <mo>=</mo> <mn>2048</mn> </mrow> </semantics></math>).</p>
Full article ">Figure 1 Cont.
<p>Spectrogram difference between original data and stego data, with a red mark on the borderline caused by the watermarking: Ja_m5.wav (<math display="inline"><semantics> <mrow> <mi>N</mi> <mo>=</mo> <mn>2048</mn> </mrow> </semantics></math>).</p>
Full article ">Figure 2
<p>Comparison of stego data generated from our previous work: random expansion (solution 1) and adaptive expansion (solution 2) with different expansion locations. Top: DCT coefficients of the original (cover) data. Middle: DCT coefficients of the stego data. Bottom: Differences between cover and stego data.</p>
Full article ">Figure 3
<p>Comparison of the audio quality of our previous work (conventional), random expansion (solution 1), and adaptive expansion (solution 2) with different expansion locations.</p>
Full article ">Figure 3 Cont.
<p>Comparison of the audio quality of our previous work (conventional), random expansion (solution 1), and adaptive expansion (solution 2) with different expansion locations.</p>
Full article ">Figure 4
<p>Comparison of the spectrum between stego data generated by the method without adaption and with adaption (64 segments): Ja_m5.wav.</p>
Full article ">Figure 5
<p>Examples of the discontinuity values for the cover and stego data.</p>
Full article ">Figure 5 Cont.
<p>Examples of the discontinuity values for the cover and stego data.</p>
Full article ">Figure 6
<p>FAR and FRR values of the proposed method to determine the EER threshold.</p>
Full article ">Figure 7
<p>Computational time per 1 s speech using the conventional and proposed methods.</p>
Full article ">
22 pages, 11189 KiB  
Article
VUF-MIWS: A Visible and User-Friendly Watermarking Scheme for Medical Images
by Chia-Chen Lin, Yen-Heng Lin, En-Ting Chu, Wei-Liang Tai and Chun-Jung Lin
Electronics 2025, 14(1), 122; https://doi.org/10.3390/electronics14010122 - 30 Dec 2024
Viewed by 510
Abstract
The integration of Internet of Medical Things (IoMT) technology has revolutionized healthcare, allowing rapid access to medical images and enhancing remote diagnostics in telemedicine. However, this advancement raises serious cybersecurity concerns, particularly regarding unauthorized access and data integrity. This paper presents a novel, [...] Read more.
The integration of Internet of Medical Things (IoMT) technology has revolutionized healthcare, allowing rapid access to medical images and enhancing remote diagnostics in telemedicine. However, this advancement raises serious cybersecurity concerns, particularly regarding unauthorized access and data integrity. This paper presents a novel, user-friendly, visible watermarking scheme for medical images—Visual and User-Friendly Medical Image Watermarking Scheme (VUF-MIWS)—designed to secure medical image ownership while maintaining usability for diagnostic purposes. VUF-MIWS employs a unique combination of inpainting and data hiding techniques to embed hospital logos as visible watermarks, which can be removed seamlessly once image authenticity is verified, restoring the image to its original state. Experimental results demonstrate the scheme’s robust performance, with the watermarking process preserving critical diagnostic information with high fidelity. The method achieved Peak Signal-to-Noise Ratios (PSNR) above 70 dB and Structural Similarity Index Measures (SSIM) of 0.99 for inpainted images, indicating minimal loss of image quality. Additionally, VUF-MIWS effectively restored the ROI region of medical images post-watermark removal, as verified through test cases with restored watermarked regions matching the original images. These findings affirm VUF-MIWS’s suitability for secure telemedicine applications. Full article
Show Figures

Figure 1

Figure 1
<p>Extra verification procedure for doctors.</p>
Full article ">Figure 2
<p>Framework of Yu et al.’s enhanced generative inpainting framework [<a href="#B19-electronics-14-00122" class="html-bibr">19</a>].</p>
Full article ">Figure 3
<p>Inpainting results of Yu et al.’s scheme [<a href="#B19-electronics-14-00122" class="html-bibr">19</a>]. (<b>a</b>) Original image; (<b>b</b>) Original image; (<b>c</b>) Original image; (<b>d</b>) Image (<b>a</b>) with a mask; (<b>e</b>) Image (<b>b</b>) with a mask; (<b>f</b>) Image (<b>c</b>) with a mask; (<b>g</b>) Inpainting results of (<b>d</b>) (PSNR = 26.45 dB); (<b>h</b>) Inpainting results of (<b>e</b>) (PSNR = 43.37 dB); (<b>i</b>) Inpainting results of (<b>f</b>) (PSNR = 54.21 dB).</p>
Full article ">Figure 4
<p>The enhancement of Yu et al.’s [<a href="#B19-electronics-14-00122" class="html-bibr">19</a>] model and the details of GAN network. (<b>a</b>) The enhancement of Yu et al.’s [<a href="#B19-electronics-14-00122" class="html-bibr">19</a>] model; (<b>b</b>) Generative Adversarial Network.</p>
Full article ">Figure 5
<p>Framework of the proposed VUF-MIWS.</p>
Full article ">Figure 6
<p>Flowchart of the recovery information generation phase.</p>
Full article ">Figure 7
<p>Flowchart of the embedding phase.</p>
Full article ">Figure 8
<p>The circular hiding path for embedding at the LL subband.</p>
Full article ">Figure 9
<p>Flowchart of the watermark removal and restoration.</p>
Full article ">Figure 10
<p>Eight medical test images. (<b>a</b>) 10.png; (<b>b</b>) 11.png; (<b>c</b>) 14.png; (<b>d</b>) 16.png; (<b>e</b>) 19.png; (<b>f</b>) 26.png; (<b>g</b>) 31.png; (<b>h</b>) 57.png.</p>
Full article ">Figure 11
<p>Two datasets are used to test the stable performance of the proposed scheme. (<b>a</b>–<b>d</b>) are Dataset 1, images of the pituitary gland taken from back to front. (<b>e</b>–<b>h</b>) are Dataset 2, images of the pituitary gland taken from top to bottom.</p>
Full article ">Figure 12
<p>Six general grayscale images sized 512 × 512 are used for the third experiment.</p>
Full article ">Figure 13
<p>The logo with a size of 64 × 64. (<b>a</b>) NCUT logo, and (<b>b</b>) Squirrel logo.</p>
Full article ">Figure 14
<p>In the first and second experiments, nine sub-regions were designated as position candidates for the visible watermark.</p>
Full article ">Figure 15
<p>Eight watermarked images. (<b>a</b>) Watermarked 10.png; (<b>b</b>) Watermarked 11.png; (<b>c</b>) Watermarked 14.png; (<b>d</b>) Watermarked 16.png; (<b>e</b>) Watermarked 19.png; (<b>f</b>) Watermarked 26.png; (<b>g</b>) Watermarked 31.png; (<b>h</b>) Watermarked 57.png.</p>
Full article ">Figure 16
<p>The restored images. (<b>a</b>) Restored 10.png; (<b>b</b>) Restored 11.png; (<b>c</b>) Restored 14.png; (<b>d</b>) Restored 16.png; (<b>e</b>) Restored 19.png; (<b>f</b>) Restored 26.png; (<b>g</b>) Restored 31.png; (<b>h</b>) Restored 57.png.</p>
Full article ">Figure 17
<p>Image recovery analysis. (<b>a</b>) Enlarged part of 11.png; (<b>b</b>) Enlarged watermarked of (<b>a</b>); (<b>c</b>) Restored image of (<b>b</b>); (<b>d</b>) Enlarged part of 14.png; (<b>e</b>) Enlarged watermarked of (<b>d</b>); (<b>f</b>) Restored image of (<b>e</b>).</p>
Full article ">Figure 18
<p>Inpainting results analysis. (<b>a</b>) Inpainting results of Yu [<a href="#B19-electronics-14-00122" class="html-bibr">19</a>]; (<b>b</b>) Histogram analysis of Yu [<a href="#B19-electronics-14-00122" class="html-bibr">19</a>]; (<b>c</b>) Inpainting results of the proposed scheme; (<b>d</b>) Histogram analysis of the proposed scheme.</p>
Full article ">
24 pages, 3170 KiB  
Article
AGFI-GAN: An Attention-Guided and Feature-Integrated Watermarking Model Based on Generative Adversarial Network Framework for Secure and Auditable Medical Imaging Application
by Xinyun Liu, Ronghua Xu and Chen Zhao
Electronics 2025, 14(1), 86; https://doi.org/10.3390/electronics14010086 - 28 Dec 2024
Viewed by 644
Abstract
With the rapid digitization of healthcare, the secure transmission of medical images has become a critical concern, especially given the increasing prevalence of cyber threats and data privacy breaches. Medical images are frequently transmitted via the Internet and cloud platforms, making them susceptible [...] Read more.
With the rapid digitization of healthcare, the secure transmission of medical images has become a critical concern, especially given the increasing prevalence of cyber threats and data privacy breaches. Medical images are frequently transmitted via the Internet and cloud platforms, making them susceptible to unauthorized access, tampering, and theft. While traditional cryptographic techniques play a vital role, they are often insufficient to fully ensure the integrity and confidentiality of these sensitive images. In this paper, we present AGFI-GAN, a robust and secure framework for medical image watermarking that leverages attention-guided and Feature-Integrated mechanisms within a Generative Adversarial Network (GAN). Specifically, a Feature-Integrated Module (FIM) is proposed to effectively capture and combine both shallow and deep image features to facilitate multilayer fusion with the watermark. The dense connections within the module facilitate feature reuse, boosting the system’s robustness. To mitigate distortion from watermark embedding, an Attention Module (AM) is utilized, generating an attention mask by extracting global image features. This attention mask prioritizes features in less prominent and textured regions, allowing for stronger watermark embedding, while other features are downplayed to enhance the overall effectiveness of the watermarking process. The framework is evaluated based on its versatility, embedding capacity, robustness, and imperceptibility, and the results confirm its effectiveness. The study shows a marked improvement over the baseline, thus highlighting the framework’s superiority. Full article
(This article belongs to the Special Issue AI Synergy: Vision, Language, and Modality)
Show Figures

Figure 1

Figure 1
<p>Residual network. The X represents the output value from the previous layer that is fed into the neuron.</p>
Full article ">Figure 2
<p>The DenseNet architecture captures both shallow and deep features, which are subsequently combined with the watermark to strengthen its resilience.</p>
Full article ">Figure 3
<p>Architecture of attention-guided module, which employs both channel attention and spatial attention to enhance the performance of image feature extraction. It helps embed the watermark in less noticeable regions.</p>
Full article ">Figure 4
<p>The architecture of the proposed AGFI-GAN is designed as an end-to-end watermarking framework capable of automatically producing watermarks that are both imperceptible and resilient. Its main components consist of an encoder, decoder, noise subnetwork, and a discriminator.</p>
Full article ">Figure 5
<p>Encoder architecture. The encoder consists of (1) a feature-integrated module (FIM) that leverages dense connections to extract both shallow and deep features, which are subsequently combined with the watermark to enhance its robustness, and (2) an attention-guided module (AGM) that uses both spatial attention and channel attention to embed the watermark in less perceptible areas of the original image.</p>
Full article ">Figure 6
<p>Performance of watermarking in AGFI-GAN. (<b>a</b>) Original image. (<b>b</b>) Watermarked image. The original and watermarked images appear nearly identical, demonstrating that the watermark has been effectively embedded in imperceptible regions.</p>
Full article ">Figure 7
<p>Subjective assessment by comparing the histograms of the original and watermarked images.</p>
Full article ">Figure 8
<p>The result of SIFT feature matching between the original and watermarked images. The colorful lines are used to link the matching keypoints between two images.</p>
Full article ">Figure 9
<p>Comparative analysis of proposed work with HiDDeN [<a href="#B27-electronics-14-00086" class="html-bibr">27</a>], TSDL [<a href="#B48-electronics-14-00086" class="html-bibr">48</a>], MBRS [<a href="#B49-electronics-14-00086" class="html-bibr">49</a>], ReDMark [<a href="#B16-electronics-14-00086" class="html-bibr">16</a>]. The vertical axis represents bit accuracy.</p>
Full article ">Figure 10
<p>The impact of watermarking on the classification accuracy of medical images.</p>
Full article ">
Back to TopTop