Digital Audio Tampering Detection Based on Deep Temporal–Spatial Features of Electrical Network Frequency
<p>Digital audio tampering detection task flowchart, where <math display="inline"><semantics> <mrow> <mi>S</mi> <mi>c</mi> <mi>o</mi> <mi>r</mi> <mi>e</mi> <mn>1</mn> </mrow> </semantics></math> and <math display="inline"><semantics> <mrow> <mi>S</mi> <mi>c</mi> <mi>o</mi> <mi>r</mi> <mi>e</mi> <mn>2</mn> </mrow> </semantics></math> denote the probabilities of tampered and untampered categories and <math display="inline"><semantics> <msup> <mi>S</mi> <mo>*</mo> </msup> </semantics></math> denotes the maximum of <math display="inline"><semantics> <mrow> <mi>S</mi> <mi>c</mi> <mi>o</mi> <mi>r</mi> <mi>e</mi> <mn>1</mn> </mrow> </semantics></math> and <math display="inline"><semantics> <mrow> <mi>S</mi> <mi>c</mi> <mi>o</mi> <mi>r</mi> <mi>e</mi> <mn>2</mn> </mrow> </semantics></math>.</p> "> Figure 2
<p>A framework diagram of digital audio tampering detection based on ENF deep temporal–spatial feature, the model is divided into two steps: (1) shallow temporal and spatial feature extraction and (2) parallel RDTCN-CNN network model construction.</p> "> Figure 3
<p>Phase curve diagramof the original and tampered audio. (<b>a</b>) is the waveform graph of the original audio, (<b>b</b>) is the waveform graph of the tampered audio. When the audio is not tampered with, the phase curve is relatively smooth, as shown in (<b>c</b>). The phase of the audio; when phase tampering occurs, the phase curve changes abruptly at the point of tampering, as shown in (<b>d</b>). The phase of the tampered audio, where the audio undergoes a tampering operation of deletion at around 9 s.</p> "> Figure 4
<p>RDTCN network structure figure (<span class="html-italic">l</span>: activation values in the <span class="html-italic">l</span>-th layer, <span class="html-italic">d</span>: dilation rate, +: concatenate operation, ⊕: add operation).</p> "> Figure 5
<p>Based on the CNN deep space feature extraction module, the network input is the shallow space features <math display="inline"><semantics> <msub> <mi>S</mi> <mrow> <mi>m</mi> <mo>×</mo> <mi>m</mi> </mrow> </msub> </semantics></math>, <span class="html-italic">m</span> is 45, and the output is the deep space features extracted by the CNN network.</p> "> Figure 6
<p>Branch attention mechanism, where ⊕ is the <math display="inline"><semantics> <mrow> <mi>c</mi> <mi>o</mi> <mi>n</mi> <mi>c</mi> <mi>a</mi> <mi>t</mi> </mrow> </semantics></math> operation and ⊗ is the dot product.</p> "> Figure 7
<p>Comparison between this method and the four baseline methods under different datasets and different evaluation indexes, respectively. DFT1-SVM, ES-SVM, PF-SVM, and X-BiLSTM are the baseline methods, RDTCN-CNN is this paper’s method, <math display="inline"><semantics> <mrow> <mi>A</mi> <mi>C</mi> <mi>C</mi> </mrow> </semantics></math> and <math display="inline"><semantics> <mrow> <mi>F</mi> <mn>1</mn> <mtext>-</mtext> <mi>s</mi> <mi>c</mi> <mi>o</mi> <mi>r</mi> <mi>e</mi> </mrow> </semantics></math> are the evaluation metrics, and Carioca, New Spanish, and ENF_Audio are the three audio tampering detection databases.</p> "> Figure 8
<p>Comparison between this method and the four baseline methods under different datasets and different evaluation indexes, respectively.</p> ">
Abstract
:1. Introduction
- Based on the extraction of high-precision ENF phase sequences, we frame the ENF according to its temporal volatility variation to represent the temporal features of the ENF and frame the ENF by adaptive frameshifting to obtain a phase feature matrix of the same size to represent the spatial features of the ENF. The feature representation capability is enhanced by deeply mining the tampering information of different dimensions in the ENF.
- We exploit the excellent modeling ability of RDTCN in the time domain and the spatial representation ability of CNN to extract deep temporal and spatial features and use the branch attention mechanism to adaptively assign the weights of temporal and spatial information to achieve the fusion of temporal–spatial features. The fused temporal–spatial features, with complementary advantages, are beneficial to improve detection accuracy. The implementation code for this study is posted at https://github.com/CCNUZFW/DTSF-ENF (accessed on 21 February 2023).
- The proposed framework achieves state-of-the-art performance on the datasets Carioca, New Spanish, and ENF_Audio compared to the four baseline methods. Compared with the baseline model, the accuracy is improved by 0.80% to 7.51% and the F1-score is improved by 0.86% to 7.53%.
2. Related Work
2.1. Based on Background Noise Consistency Detection
2.2. Based on the Analysis of Audio Content Features
2.3. Based on Electrical Network Frequency Consistency Detection
3. Preliminaries
3.1. Problem Definition
3.2. ENF Shallow Temporal and Spatial Feature Definition
3.3. ENF Deep Temporal and Spatial Feature Definition
4. Methods
4.1. Extraction of the Shallow Temporal and Spatial Features of ENF
4.1.1. Extraction of First-Order Phase Features
Algorithm 1 Extraction of first-order phase features . |
|
4.1.2. Extraction of ENF Shallow Temporal Features
Algorithm 2 Temporal feature frame processing of ENF. |
|
4.1.3. Extraction of ENF Shallow Spatial Features
Algorithm 3 Spatial feature frame processing of ENF. |
|
4.2. Deep Feature Representation Learning Based on RDTCN-CNN Temporal–Spatial Feature Fusion
4.2.1. Deep Temporal Feature Extraction Based on RDTCN Network
4.2.2. Deep Spatial Feature Extraction Based on CNN Network
4.2.3. Deep Temporal and Spatial Feature Fusion Based on Branch Attention Mechanism
4.2.4. Classification Network Design
5. Experimental Results and Analysis
5.1. Dataset
5.2. Evaluation Metrics
5.3. Baselines
5.4. Experimental Settings
5.5. Results and Discussion
5.5.1. Comparison with Baseline Methods
5.5.2. Verifying the Effectiveness of the RDTCN Temporal Feature Extraction Network
5.5.3. Verifying the Effect of Frame Length Setting on the Shallow Temporal Features of ENF
5.5.4. Verifying the Effectiveness of the Branch Attention Mechanism
6. Conclusions
Author Contributions
Funding
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
Abbreviations
ENF | Electrical Network Frequency |
RDTCN | Residual Dense Temporal Convolutional Networks |
CNN | Convolutional Neural Networks |
MLP | Multilayer Perceptron |
MFCC | Mel Frequency Cepstral Coefficient |
SVM | Support Vector Machine |
ENFC | ENF Component |
RFA | Robust Filtering Algorithm |
PSTN | Public Switched Telephone Network |
RNN | Recurrent Neural Networks |
LSTM | Long Short-Term Memory |
References
- Liu, Z.; Lu, W. Fast Copy-Move Detection of Digital Audio. In Proceedings of the 2017 IEEE Second International Conference on Data Science in Cyberspace (DSC), Shenzhen, China, 26–29 June 2017; pp. 625–629. [Google Scholar] [CrossRef]
- Zeng, C.; Zhu, D.; Wang, Z.; Wang, Z.; Zhao, N.; He, L. An End-to-End Deep Source Recording Device Identification System for Web Media Forensics. Int. J. Web Inf. Syst. 2020, 16, 413–425. [Google Scholar] [CrossRef]
- Yan, Q.; Yang, R.; Huang, J. Detection of Speech Smoothing on Very Short Clip. IEEE Trans. Inf. Forensics Secur. 2019, 9, 2441–2453. [Google Scholar] [CrossRef]
- Wang, Z.; Yang, Y.; Zeng, C.; Kong, S.; Feng, S.; Zhao, N. Shallow and Deep Feature Fusion for Digital Audio Tampering Detection. EURASIP J. Adv. Signal Process. 2022, 2022, 1–20. [Google Scholar] [CrossRef]
- Zeng, C.; Yang, Y.; Wang, Z.; Kong, S.; Feng, S. Audio Tampering Forensics Based on Representation Learning of ENF Phase Sequence. Int. J. Digit. Crime Forensics 2022, 14, 1–19. [Google Scholar] [CrossRef]
- Wang, Z.F.; Wang, J.; Zeng, C.Y.; Min, Q.S.; Tian, Y.; Zuo, M.Z. Digital Audio Tampering Detection Based on ENF Consistency. In Proceedings of the 2018 International Conference on Wavelet Analysis and Pattern Recognition (ICWAPR) IEEE, Chengdu, China, 15–18 July 2018; pp. 209–214. [Google Scholar] [CrossRef]
- Hua, G.; Liao, H.; Wang, Q. Detection of Electric Network Frequency in Audio Recordings–From Theory to Practical Detectors; IEEE Press: Piscataway, NJ, USA, 2021; Volume 1, pp. 1556–6013. [Google Scholar] [CrossRef]
- Hajj-Ahmad, A.; Garg, R.; Wu, M. Instantaneous frequency estimation and localization for ENF signals. In Proceedings of the 2012 Asia Pacific Signal and Information Processing Association Annual Summit and Conference IEEE, Hollywood, CA, USA, 3–6 December 2012; pp. 1–10. [Google Scholar]
- Bykhovsky, D. Recording Device Identification by ENF Harmonics Power Analysis. Forensic Sci. Int. 2020, 307, 110100. [Google Scholar] [CrossRef] [PubMed]
- Zeng, C.; Zhu, D.; Wang, Z.; Wu, M.; Xiong, W.; Zhao, N. Spatial and Temporal Learning Representation for End-to-End Recording Device Identification. EURASIP J. Adv. Signal Process. 2021, 2021, 41. [Google Scholar] [CrossRef]
- Lin, X.; Zhu, J.; Chen, D. Subband Aware CNN for Cell-Phone Recognition. IEEE Signal Process. Lett. 2020, 27, 5. [Google Scholar] [CrossRef]
- Verma, V.; Khanna, N. Speaker-Independent Source Cell-Phone Identification for Re-Compressed and Noisy Audio Recordings. Multimed. Tools Appl. 2021, 80, 23581–23603. [Google Scholar] [CrossRef]
- Meng, X.; Li, C.; Tian, L. Detecting Audio Splicing Forgery Algorithm Based on Local Noise Level Estimation. In Proceedings of the 2018 5th International Conference on Systems and Informatics (ICSAI), Nanjing, China, 10–12 November 2018; pp. 861–865. [Google Scholar] [CrossRef]
- Lin, X.; Kang, X. Exposing speech tampering via spectral phase analysis. Digit. Signal Process. 2017, 1, 63–74. [Google Scholar] [CrossRef]
- Yan, D.; Dong, M.; Gao, J. Exposing Speech Transsplicing Forgery with Noise Level Inconsistency. Secur. Commun. Netw. 2021, 1, 6. [Google Scholar] [CrossRef]
- Narkhede, M.; Patole, R. Acoustic scene identification for audio authentication. Soft Comput. Signal Process. 2021, 1, 593–602. [Google Scholar]
- Capoferri, D.; Borrelli, C. Speech Audio Splicing Detection and Localization Exploiting Reverberation Cues. In Proceedings of the 2020 IEEE International Workshop on Information Forensics and Security (WIFS), New York, NY, USA, 6–11 December 2020; pp. 1–6. [Google Scholar] [CrossRef]
- Jadhav, S.; Patole, R.; Rege, P. Audio Splicing Detection using Convolutional Neural Network. In Proceedings of the 2019 10th International Conference on Computing, Communication and Networking Technologies (ICCCNT), Kanpur, India, 6–8 July 2019; pp. 1–5. [Google Scholar] [CrossRef]
- Saleem, S.; Dilawari, A.; Khan, U. Spoofed Voice Detection using Dense Features of STFT and MDCT Spectrograms. In Proceedings of the 2021 International Conference on Artificial Intelligence (ICAI), Islamabad, Pakistan, 5–7 April 2021; pp. 56–61. [Google Scholar] [CrossRef]
- Li, C.; Sun, Y.; Meng, X. Homologous Audio Copy-move Tampering Detection Method Based on Pitch. In Proceedings of the 2019 IEEE 19th International Conference on Communication Technology (ICCT), Xi’an, China, 16–19 October 2019; pp. 530–534. [Google Scholar] [CrossRef]
- Yan, Q.; Yang, R.; Huang, J. Robust Copy–Move Detection of Speech Recording Using Similarities of Pitch and Formant. IEEE Trans. Inf. Forensics Secur. 2019, 9, 2331–2341. [Google Scholar] [CrossRef]
- Xie, X.; Lu, W.; Liu, X. Copy-move detection of digital audio based on multi-feature decision. J. Inf. Secur. Appl. 2018, 10, 37–46. [Google Scholar] [CrossRef]
- Lin, X.; Kang, X. Supervised audio tampering detection using an autoregressive model. In Proceedings of the 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), New Orleans, LA, USA, 5–9 March 2017; pp. 2142–2146. [Google Scholar] [CrossRef]
- Hua, G.; Liao, H.; Zhang, H. Robust ENF Estimation Based on Harmonic Enhancement and Maximum Weight Clique. IEEE Trans. Inf. Forensics Secur. 2021, 7, 3874–3887. [Google Scholar] [CrossRef]
- Nicolalde, D.; Apolinario, J. Audio Authenticity: Detecting ENF Discontinuity With High Precision Phase Analysis. IEEE Trans. Inf. Forensics Secur. 2010, 9, 534–543. [Google Scholar] [CrossRef]
- Reis, P.; Lustosa, J.; Miranda, R. ESPRIT-Hilbert-Based Audio Tampering Detection With SVM Classifier for Forensic Analysis via Electrical Network Frequency. IEEE Trans. Inf. Forensics Secur. 2017, 4, 853–864. [Google Scholar] [CrossRef]
- Zakariah, M.; Khan, M.; Malik, H. Digital multimedia audio forensics: Past, present and future. Multimed. Tools Appl. 2017, 1, 1009–1040. [Google Scholar] [CrossRef]
- Bai, Z.; Zhang, X.L. Speaker Recognition Based on Deep Learning: An Overview. Neural Netw. 2021, 140, 65–99. [Google Scholar] [CrossRef]
- Mohd Hanifa, R.; Isa, K.; Mohamad, S. A Review on Speaker Recognition: Technology and Challenges. Comput. Electr. Eng. 2021, 90, 107005. [Google Scholar] [CrossRef]
- Wang, Z.; Wang, Z.; Zeng, C.; Yu, Y.; Wan, X. High-Quality Image Compressed Sensing and Reconstruction with Multi-Scale Dilated Convolutional Neural Network. Circuits Syst. Signal Process. 2022, 42, 1–24. [Google Scholar] [CrossRef]
- Abdu, S.A.; Yousef, A.H.; Salem, A. Multimodal Video Sentiment Analysis Using Deep Learning Approaches, a Survey. Inf. Fusion 2021, 76, 204–226. [Google Scholar] [CrossRef]
- Bayoudh, K.; Knani, R.; Hamdaoui, F.; Mtibaa, A. A Survey on Deep Multimodal Learning for Computer Vision: Advances, Trends, Applications, and Datasets. Vis. Comput. 2022, 38, 2939–2970. [Google Scholar] [CrossRef]
- Chango, W.; Lara, J.A.; Cerezo, R.; Romero, C. A Review on Data Fusion in Multimodal Learning Analytics and Educational Data Mining. WIREs Data Min. Knowl. Discov. 2022, 12, e1458. [Google Scholar] [CrossRef]
- Dimitri, G.M. A Short Survey on Deep Learning for Multimodal Integration: Applications, Future Perspectives and Challenges. Computers 2022, 11, 163. [Google Scholar] [CrossRef]
- Gandhi, A.; Adhvaryu, K.; Poria, S.; Cambria, E.; Hussain, A. Multimodal Sentiment Analysis: A Systematic Review of History, Datasets, Multimodal Fusion Methods, Applications, Challenges and Future Directions. Inf. Fusion 2023, 91, 424–444. [Google Scholar] [CrossRef]
- Han, X.; Wang, Y.T.; Feng, J.L.; Deng, C.; Chen, Z.H.; Huang, Y.A.; Su, H.; Hu, L.; Hu, P.W. A Survey of Transformer-Based Multimodal Pre-Trained Modals. Neurocomputing 2023, 515, 89–106. [Google Scholar] [CrossRef]
- He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
- Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet Classification with Deep Convolutional Neural Networks. In Proceedings of the Advances in Neural Information Processing Systems, Lake Tahoe, NV, USA, 3–6 December 2012; Volume 25. [Google Scholar]
- Wang, Z.; Yan, W.; Zeng, C.; Tian, Y.; Dong, S. A Unified Interpretable Intelligent Learning Diagnosis Framework for Learning Performance Prediction in Intelligent Tutoring Systems. Int. J. Intell. Syst. 2023, 2023, 1–20. [Google Scholar] [CrossRef]
- Wu, T.; Ling, Q. Self-Supervised Heterogeneous Hypergraph Network for Knowledge Tracing. Inf. Sci. 2023, 624, 200–216. [Google Scholar] [CrossRef]
- Pan, X.; Zhang, X. Detecting splicing in digital audios using local noise level estimation. In Proceedings of the 2012 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Kyoto, Japan, 25–30 March 2012; pp. 1841–1844. [Google Scholar] [CrossRef]
- Malik, H. Acoustic environment identification and its applications to audio forensics. IEEE Trans. Inf. Forensics Secur. 2013, 8, 1827–1837. [Google Scholar] [CrossRef]
- Mascia, M.; Canclini, A.; Antonacci, F. Forensic and anti-forensic analysis of indoor/outdoor classifiers based on acoustic clues. In Proceedings of the 2015 23rd European Signal Processing Conference (EUSIPCO), Nice, France, 31 August–4 September 2015; pp. 2072–2076. [Google Scholar] [CrossRef]
- Ikram, S.; Malik, H. Digital audio forensics using background noise. In Proceedings of the 2010 IEEE International Conference on Multimedia and Expo, Singapore, 19–23 July 2010; pp. 106–110. [Google Scholar] [CrossRef]
- Chen, J.; Xiang, S.; Huang, H. Detecting and locating digital audio forgeries based on singularity analysis with wavelet packet. Multimed. Tools Appl. 2016, 2, 2303–2325. [Google Scholar] [CrossRef]
- Imran, M.; Xiang, S.; Huang, H. Blind detection of copy-move forgery in digital audio forensics. IEEE Access 2017, 6, 12843–12855. [Google Scholar] [CrossRef]
- Esquef, P.A.A.; Apolinário, J.A.; Biscainho, L.W.P. Edit Detection in Speech Recordings via Instantaneous Electric Network Frequency Variations. IEEE Trans. Inf. Forensics Secur. 2014, 10, 2314–2326. [Google Scholar] [CrossRef]
- Mao, M.; Xiao, Z.; Kang, X.; Li, X. Electric Network Frequency Based Audio Forensics Using Convolutional Neural Networks. IFIP Adv. Inf. Commun. Technol. 2020, 8, 253–270. [Google Scholar] [CrossRef]
- Sarkar, M.; Chowdhury, D.; Shahnaz, C.; Fattah, S.A. Application of Electrical Network Frequency of Digital Recordings for Location-Stamp Verification. Appl. Sci. 2019, 9, 3135. [Google Scholar] [CrossRef]
- Karantaidis, G.; Kotropoulos, C. Blackman–Tukey spectral estimation and electric network frequency matching from power mains and speech recordings. IET Signal Process. 2021, 6, 396–409. [Google Scholar] [CrossRef]
- Hua, G.; Zhang, H. ENF Signal Enhancement in Audio Recordings. IEEE Trans. Inf. Forensics Secur. 2020, 11, 1868–1878. [Google Scholar] [CrossRef]
- Ortega-Garcia, J.; Gonzalez-Rodriguez, J. Audio Speech variability in automatic speaker recognition systems for commercial and forensic purposes. IEEE Aerosp. Electron. Syst. Mag. 2000, 11, 27–32. [Google Scholar] [CrossRef]
Notations | Descriptions |
---|---|
, , , n | Digital audio signal, downsampled signal, ENFC signal, n is indexed |
Downsampling frequency | |
, | 1- order , k- order |
, | Signal after of and |
k, | Indexing of signal and signal peak points per frame |
, | 0th order phase sequence, 1st order phase sequence |
Frequency value of the first-order ENF signal | |
Shallow temporal feature of ENF | |
Shallow spatial feature of ENF | |
, | Rounding down, rounding up |
Calculation formula of the frameshift | |
Dilated convolution formula | |
Binary cross-entropy loss function | |
, | Prediction accuracy, F1-score |
The Dataset | Carioca | New Spanish | ENF_Audio |
---|---|---|---|
Edited audio | 250 | 502 | 752 |
Original audio | 250 | 251 | 501 |
Total audio | 500 | 753 | 1253 |
Audio duration | 9∼35 s | 16∼35 s | 9∼35 s |
The training set | 350 | 527 | 877 |
The validation set | 50 | 75 | 125 |
The test set | 100 | 151 | 251 |
Method | Carioca | New Spanish | ENF_Audio | |||
---|---|---|---|---|---|---|
ACC (%) | F1-Score (%) | ACC (%) | F1-Score (%) | ACC (%) | F1-Score (%) | |
DFT1-SVM [25] | 89.90 | 90.22 | 88.86 | 86.84 | 90.51 | 90.55 |
ES-SVM [26] | 90.88 | 90.62 | 90.62 | 88.26 | 93.52 | 93.44 |
PF-SVM [6] | 93.05 | 92.86 | 90.22 | 87.56 | 92.60 | 92.82 |
X-BiLSTM [5] | 97.03 | 97.22 | 92.14 | 90.62 | 97.22 | 97.02 |
RDTCN-CNN | 97.96 | 97.54 | 95.60 | 94.50 | 98.02 | 97.88 |
Method | Carioca | New Spanish | ENF_Audio | |||
---|---|---|---|---|---|---|
ACC (%) | F1-Score (%) | ACC (%) | F1-Score (%) | ACC (%) | F1-Score (%) | |
Ordinary TCN | 96.58 | 96.22 | 93.56 | 91.88 | 97.42 | 97.46 |
RDTCN | 97.96 | 97.54 | 95.60 | 94.50 | 98.02 | 97.88 |
Method | Carioca | New Spanish | ENF_Audio | |||
---|---|---|---|---|---|---|
ACC (%) | F1-Score (%) | ACC (%) | F1-Score (%) | ACC (%) | F1-Score (%) | |
Splice Fusion | 96.02 | 96.42 | 94.42 | 92.82 | 97.20 | 97.22 |
Branch | 97.96 | 97.54 | 95.60 | 94.50 | 98.02 | 97.88 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Zeng, C.; Kong, S.; Wang, Z.; Li, K.; Zhao, Y. Digital Audio Tampering Detection Based on Deep Temporal–Spatial Features of Electrical Network Frequency. Information 2023, 14, 253. https://doi.org/10.3390/info14050253
Zeng C, Kong S, Wang Z, Li K, Zhao Y. Digital Audio Tampering Detection Based on Deep Temporal–Spatial Features of Electrical Network Frequency. Information. 2023; 14(5):253. https://doi.org/10.3390/info14050253
Chicago/Turabian StyleZeng, Chunyan, Shuai Kong, Zhifeng Wang, Kun Li, and Yuhao Zhao. 2023. "Digital Audio Tampering Detection Based on Deep Temporal–Spatial Features of Electrical Network Frequency" Information 14, no. 5: 253. https://doi.org/10.3390/info14050253
APA StyleZeng, C., Kong, S., Wang, Z., Li, K., & Zhao, Y. (2023). Digital Audio Tampering Detection Based on Deep Temporal–Spatial Features of Electrical Network Frequency. Information, 14(5), 253. https://doi.org/10.3390/info14050253