Frame Identification of Object-Based Video Tampering Using Symmetrically Overlapped Motion Residual
<p>Example of object-based video tampering. (<b>a</b>) Pristine frames. (<b>b</b>) Partially tampered-with frames. The object surrounded by the red ellipse is removed from the third and fourth frames.</p> "> Figure 2
<p>Two types of frame identification in object-based video tampering. (<b>a</b>) Two-class identification. (<b>b</b>) Three-class identification.</p> "> Figure 3
<p>Frame structure for collusion operation.</p> "> Figure 4
<p>Example of the median collusion operation at pixel level.</p> "> Figure 5
<p>Frame structure for overlapped motion residual generation.</p> "> Figure 6
<p>Sample motion residual images for pristine and corresponding tampered-with frame. (<b>a</b>) Conventional motion residual and overlapped three motion residual images for pristine frame; (<b>b</b>) Conventional motion residual and overlapped three motion residual images for tampered-with frame. All motion residual images are rescaled to a maximum value of 255.</p> "> Figure 7
<p>Basic convolutional neural network (CNN) structure.</p> "> Figure 8
<p>Configuration of machine learning network. (<b>a</b>) Basic network for classifying forged and non-forged patches; (<b>b</b>) Basic CNN for classifying pristine and double-compressed patches.</p> "> Figure 9
<p>Classification method explained using a logic circuit concept.</p> ">
Abstract
:1. Introduction
- (1)
- A symmetrically overlapped motion residual is proposed to enhance the discernment of video frames. Three different motion residuals are presented on the basis of overlapping temporal frames. The proposed overlapped motion residuals can reduce natural motions while preserving motions caused by tampering. Because the overlapped motion residuals use three different temporal windows, temporal changes in motion residuals can be exploited in the neural network.
- (2)
- An asymmetric CNN structure for training and testing is proposed. In the training step, we design a single basic CNN and use it to perform two different training processes, resulting in two trained networks with an identical structure. In the testing process, we propose two types of testing methods corresponding to two- and three-class frame identifications. Two-class identification determines whether a video frame has been forged using a single trained network. In contrast, three-class identification is also performed to classify whether a frame is pristine, double-compressed, or tampered with by combining the two learned basic networks.
2. Object-Based Video Forgery and Detection
3. Symmetrically Overlapped Motion Residual
3.1. Motion Residual
3.2. Proposed Overlapped Motion Resiual
3.3. Patch Generation of Overlapped Motion Resiual
4. Proposed Frame Identification
4.1. Basic Network
4.2. Training
4.3. Classification
5. Experimental Results
5.1. Dataset
5.2. Experimental Setup
5.3. Results
5.3.1. Two-Class Identification
5.3.2. Three-Class Identification
6. Discussion
7. Conclusions
Author Contributions
Funding
Conflicts of Interest
References
- Alsmirat, M.A.; Al-Hussien, R.A.; Al-Sarayrah, W.T.; Jararweh, Y.; Etier, M. Digital video forensics: A comprehensive survey. Int. J. Adv. Intell. Paradig. 2020, 15, 437–456. [Google Scholar] [CrossRef]
- Javed, A.R.; Jalil, Z.; Zehra, W.; Gadekallu, T.R.; Suh, D.Y.; Piran, M.J. A comprehensive survey on digital video forensics: Taxonomy, challenges, and future directions. Eng. Appl. Artif. Intell. 2021, 106, 104456. [Google Scholar] [CrossRef]
- Wang, W.; Farid, H. Exposing digital forgeries in video by detecting double MPEG compression. In Proceedings of the 8th Workshop on Multimedia and Security, Geneva, Switzerland, 26 September 2006; pp. 37–47. [Google Scholar] [CrossRef]
- Vazquez-Padin, D.; Fontani, M.; Bianchi, T.; Comesana, P.; Piva, A.; Barni, M. Detection of video double encoding with GOP size estimation. In Proceedings of the IEEE International Workshop on Information Forensics and Security, Costa Adeje, Spain, 2–5 December 2012; pp. 151–156. [Google Scholar] [CrossRef] [Green Version]
- Milani, S.; Bestagini, P.; Tagliasacchi, M.; Tubaro, S. Multiple compression detection for video sequences. In Proceedings of the IEEE 14th International Workshop on Multimedia Signal Processing, Banff, AB, Canada, 17–19 September 2012; pp. 112–117. [Google Scholar] [CrossRef]
- Jiang, X.; Wang, W.; Sun, T.; Shi, Y.Q.; Wang, S. Detection of double compression in MPEG-4 videos based on Markov statistics. IEEE Signal Process. Lett. 2013, 20, 447–450. [Google Scholar] [CrossRef]
- Stamm, M.C.; Lin, W.S.; Liu, K.R. Temporal forensics and anti-forensics for motion compensated video. IEEE Trans. Inf. Forensics Secur. 2012, 7, 1315–1329. [Google Scholar] [CrossRef] [Green Version]
- Gironi, A.; Fontani, M.; Bianchi, T.; Piva, A.; Barni, M. A video forensic technique for detecting frame deletion and insertion. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing, Florence, Italy, 4–9 May 2014; pp. 6226–6230. [Google Scholar] [CrossRef] [Green Version]
- Yang, J.; Huang, T.; Su, L. Using similarity analysis to detect frame duplication forgery in videos. Multimed. Tools Appl. 2016, 75, 1793–1811. [Google Scholar] [CrossRef]
- Vaishnavi, D.; Subashini, T.S. Application of local invariant symmetry features to detect and localize image copy move forgeries. J. Inf. Secur. Appl. 2019, 44, 23–31. [Google Scholar] [CrossRef]
- Li, Y.; Zhou, J. Fast and effective image copy-move forgery detection via hierarchical feature point matching. IEEE Trans. Inf. Forensics Secur. 2019, 14, 1307–1322. [Google Scholar] [CrossRef]
- Park, J.Y.; Kang, T.A.; Moon, Y.H.; Eom, I.K. Copy-move forgery detection using scale invariant feature and reduced local binary pattern histogram. Symmetry 2020, 12, 492. [Google Scholar] [CrossRef] [Green Version]
- Ricaho, C.; Gaobo, Y.; Ningbo, Z. Detection of object-based manipulation by the statistical features of object contour. Forensic Sci. Int. 2014, 236, 164–169. [Google Scholar] [CrossRef]
- Tan, S.; Chen, S.; Li, B. GOP based automatic detection of object-based forgery in advanced video. In Proceedings of the Asia-Pacific Signal and Information Processing Association Annual Summit and Conference, Hong Kong, China, 16–19 December 2015; pp. 719–722. [Google Scholar] [CrossRef]
- Kodovsky, J.; Fridrich, J. Calibration revisited. In Proceedings of the 11th ACM workshop on Multimedia and security, Princeton, NJ, USA, 7–8 September 2009; pp. 63–74. [Google Scholar] [CrossRef]
- Kodovsky, J.; Fridrich, J. Ensemble classifiers for steganalysis of digital media. IEEE Trans. Inf. Forensics Secur. 2012, 7, 432–444. [Google Scholar] [CrossRef] [Green Version]
- Chen, S.; Tan, S.; Li, B.; Huang, J. Automatic detection of object-based forgery in advanced video. IEEE Trans. Circuits Syst. Video Technol. 2016, 26, 2138–2151. [Google Scholar] [CrossRef]
- Penvy, T.; Fridrich, J. Merging Markov and DCT features for multi-class JPEG steganalysis. In Proceedings of the SPIE Security, Steganography, and Watermarking of Multimedia Contents IX, San Jose, CA, USA, 2 March 2007; p. 650506. [Google Scholar] [CrossRef] [Green Version]
- Saddique, M.; Asghar, K.; Bajwa, U.I.; Hussain, M.; Habib, Z. Spatial video forgery detection and localization using texture analysis of consecutive frames. Adv. Electr. Comput. Eng. 2019, 19, 97–108. [Google Scholar] [CrossRef]
- Sitara, K.; Mehtre, B.M. Differentiating synthetic and optical zooming for passive video forgery detection: An anti-forensic perspective. Digit. Investig. 2019, 30, 1–11. [Google Scholar] [CrossRef]
- Aloraini, M.; Sharifzadeh, M.; Schonfeld, D. Sequential and patch analyses for object removal video forgery detection and localization. IEEE Trans. Circuits Syst. Video Technol. 2021, 31, 917–930. [Google Scholar] [CrossRef]
- Doulamis, N.; Doulamis, A.; Protopapadakis, E. Deep learning for computer vision: A brief review. Comput. Intell. Neurosci. 2018, 2018, 7068349. [Google Scholar] [CrossRef]
- Yao, Y.; Shi, Y.; Weng, S.; Guan, B. Deep learning for detection of object-based forgery in advanced video. Symmetry 2018, 10, 3. [Google Scholar] [CrossRef] [Green Version]
- Kohli, A.; Gupta, A.; Singhal, D. CNN based localisation of forged region in object-based forgery for HD videos. IET Image Process. 2020, 14, 947–958. [Google Scholar] [CrossRef]
- Yang, Q.; Yu, D.; Zhang, Z.; Yao, Y.; Chen, L. Spatiotemporal trident networks: Detection and localization of object removal tampering in video passive forensics. IEEE Trans. Circuits Syst. Video Technol. 2021, 31, 4131–4144. [Google Scholar] [CrossRef]
- Qian, Y.; Dong, J.; Wang, W.; Tan, T. Deep learning for steganalysis via convolutional neural networks. In Proceedings of the SPIE 9409, Media Watermarking, Security, and Forensics, San Francisco, CA, USA, 9–11 February 2015; p. 94090J. [Google Scholar] [CrossRef]
- Fridrich, J.; Kodovsky, J. Rich models for steganalysis of digital images. IEEE Trans. Inf. Forensics Secur. 2012, 7, 868–882. [Google Scholar] [CrossRef] [Green Version]
- Bishop, C.M. Pattern Recognition and Machine Learning; Springer: New York, NY, USA, 2006. [Google Scholar]
- Qadir, G.; Yahaya, S.; Ho, A.T.S. Surrey university library for forensic analysis (SULFA) of video content. In Proceedings of the IET Conference on Image Processing, London, UK, 3–4 July 2012; p. 94090J. [Google Scholar] [CrossRef]
- Bestagini, P.; Milani, S.; Tagliasacchi, M.; Tubaro, S. Local tampering detection in video sequences. In Proceedings of the 2013 IEEE 15th International Workshop on Multimedia Signal Processing, Pula, Italy, 30 September–2 October 2013; pp. 488–493. [Google Scholar] [CrossRef]
- Aloraini, M.; Sharifzadeh, M.; Agarwal, C.; Schonfeld, D. Statistical sequential analysis for object-based video forgery detection. Electron. Imag. 2019, 2019, 543-1–543-7. [Google Scholar] [CrossRef]
- Pevny, T.; Bas, P.; Fridrich, J. Steganalysis by subtractive pixel adjacency matrix. IEEE Trans. Inf. Forensics Secur. 2010, 5, 215–224. [Google Scholar] [CrossRef] [Green Version]
- Kodovsky, J.; Pevny, T.; Fridrich, J. Modern steganalysis can detect YASS. In Proceedings of the SPIE 7541, Media Forensics and Security II, San Jose, CA, USA, 18–20 January 2010; p. 754102. [Google Scholar] [CrossRef]
- Kodovsky, J.; Fridrich, J. Steganalysis of JPEG images using rich models. In Proceedings of the SPIE 8303, Media Watermarking, Security, and Forensics, Burlingame, CA, USA, 23–25 January 2012; p. 83030A. [Google Scholar] [CrossRef] [Green Version]
Kernel Size | Number of Kernel | Stride Size | Feature Size after AP I | |
---|---|---|---|---|
Layer I | 3 × 3 | 8 | 1 | 8 × (m’ × m’)/22 |
Layer II | 3 × 3 | 16 | 1 | 16 × (m’ × m’)/24 |
Layer III | 3 × 3 | 32 | 1 | 32 × (m’ × m’)/26 |
Layer IV | 1 × 1 | 64 | 1 | 64 × (m’ × m’)/28 |
Layer V | 1 × 1 | 128 | 1 | 128 × (m’ × m’)/210 |
AP I | 5 × 5 | 1 | 2 | - |
AP II | (m’ × m’)/210 | 1 | Global | 128 |
Actual Frame | Non-Forged | Forged | |
---|---|---|---|
Labeled Frame | |||
Non-forged | 98.10% (22,207) | 1.90% (430) | |
Forged | 1.93% (112) | 98.07% (5678) |
Method | PFACC | FFACC | TFACC | Precision | Recall | F1 Score |
---|---|---|---|---|---|---|
CNN [23] | 98.45 | 89.90 | 96.79 | 91.05 | 97.31 | 94.07 |
Temporal CNN [24] | - | 96.04 | 97.49 | - | - | - |
TF-SA [31] | - | - | - | 93.90 | 93.90 | 93.30 |
S-PA [21] | - | - | - | 95.51 | 94.44 | 94.97 |
STN [25] | 99.50 | 98.75 | 99.34 | 98.14 | 98.75 | 98.44 |
Proposed | 98.10 | 98.07 | 98.09 | 96.23 | 98.08 | 97.15 |
Actual Frame | Pristine | Double-Compressed | Forged | |
---|---|---|---|---|
Labeled Frame | ||||
Pristine | 99.89% (14,205) | 0.11% (12) | 0.00% (0) | |
Double-compressed | 0.27% (23) | 94.62% (7967) | 5.11% (430) | |
Forged | 0.00% (0) | 1.93% (112) | 98.07% (5678) |
Method | PFACC | FFACC | DFACC | TFACC | Precision | Recall | F1 Score |
---|---|---|---|---|---|---|---|
ADOBF [17]: CC + PEV [18] | 99.90% | 83.94% | 95.22% | 95.71% | 90.48% | 91.80% | 91.13% |
ADOBF [17]: SPAM [32] | 99.71% | 76.86% | 89.03% | 92.47% | 78.90% | 83.04% | 80.92% |
ADOBF [17]: CF* [33] | 99.50% | 77.55% | 93.64% | 94.15% | 87.06% | 85.87% | 86.46% |
ADOBF [17]: CDF [16] | 99.96% | 84.07% | 95.67% | 95.88% | 90.20% | 91.01% | 90.60% |
ADOBF [17]: SRM [27] | 99.91% | 76.40% | 93.21% | 93.70% | 83.10% | 82.86% | 82.89% |
ADOBF [17]: CC-FRM [34] | 99.96% | 84.93% | 97.82% | 96.59% | 93.15% | 91.51% | 92.32% |
ADOBF [17]: J + SRM [34] | 99.99% | 84.90% | 97.56% | 96.59% | 92.80% | 91.58% | 92.18% |
Proposed | 99.89% | 98.07% | 94.62% | 97.97% | 97.08% | 97.53% | 97.30% |
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2022 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Kim, T.H.; Park, C.W.; Eom, I.K. Frame Identification of Object-Based Video Tampering Using Symmetrically Overlapped Motion Residual. Symmetry 2022, 14, 364. https://doi.org/10.3390/sym14020364
Kim TH, Park CW, Eom IK. Frame Identification of Object-Based Video Tampering Using Symmetrically Overlapped Motion Residual. Symmetry. 2022; 14(2):364. https://doi.org/10.3390/sym14020364
Chicago/Turabian StyleKim, Tae Hyung, Cheol Woo Park, and Il Kyu Eom. 2022. "Frame Identification of Object-Based Video Tampering Using Symmetrically Overlapped Motion Residual" Symmetry 14, no. 2: 364. https://doi.org/10.3390/sym14020364
APA StyleKim, T. H., Park, C. W., & Eom, I. K. (2022). Frame Identification of Object-Based Video Tampering Using Symmetrically Overlapped Motion Residual. Symmetry, 14(2), 364. https://doi.org/10.3390/sym14020364