CN112699707A - Video detection method, device and storage medium - Google Patents
Video detection method, device and storage medium Download PDFInfo
- Publication number
- CN112699707A CN112699707A CN201911006641.7A CN201911006641A CN112699707A CN 112699707 A CN112699707 A CN 112699707A CN 201911006641 A CN201911006641 A CN 201911006641A CN 112699707 A CN112699707 A CN 112699707A
- Authority
- CN
- China
- Prior art keywords
- frame
- image
- features
- steganographic
- feature
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000001514 detection method Methods 0.000 title claims abstract description 46
- 230000000007 visual effect Effects 0.000 claims abstract description 121
- 238000000034 method Methods 0.000 claims abstract description 40
- 238000013527 convolutional neural network Methods 0.000 claims description 52
- 239000013598 vector Substances 0.000 claims description 52
- 238000013528 artificial neural network Methods 0.000 claims description 49
- 238000012545 processing Methods 0.000 claims description 37
- 238000000605 extraction Methods 0.000 claims description 17
- 230000000306 recurrent effect Effects 0.000 claims description 17
- 230000002159 abnormal effect Effects 0.000 claims description 10
- 230000002123 temporal effect Effects 0.000 claims description 9
- 238000011176 pooling Methods 0.000 claims description 8
- 238000012163 sequencing technique Methods 0.000 claims description 5
- 230000008569 process Effects 0.000 description 15
- 238000010586 diagram Methods 0.000 description 12
- 238000005516 engineering process Methods 0.000 description 12
- 238000004590 computer program Methods 0.000 description 11
- 238000004891 communication Methods 0.000 description 10
- 230000008859 change Effects 0.000 description 5
- 230000006870 function Effects 0.000 description 5
- 125000004122 cyclic group Chemical group 0.000 description 4
- 239000000284 extract Substances 0.000 description 4
- 238000012549 training Methods 0.000 description 4
- 102100033620 Calponin-1 Human genes 0.000 description 3
- 102100033591 Calponin-2 Human genes 0.000 description 3
- 101000945318 Homo sapiens Calponin-1 Proteins 0.000 description 3
- 101000945403 Homo sapiens Calponin-2 Proteins 0.000 description 3
- 230000005856 abnormality Effects 0.000 description 3
- 230000003287 optical effect Effects 0.000 description 3
- 230000014509 gene expression Effects 0.000 description 2
- 238000011835 investigation Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000003068 static effect Effects 0.000 description 2
- 238000012731 temporal analysis Methods 0.000 description 2
- PXFBZOLANLWPMH-UHFFFAOYSA-N 16-Epiaffinine Natural products C1C(C2=CC=CC=C2N2)=C2C(=O)CC2C(=CC)CN(C)C1C2CO PXFBZOLANLWPMH-UHFFFAOYSA-N 0.000 description 1
- 230000004075 alteration Effects 0.000 description 1
- 230000006399 behavior Effects 0.000 description 1
- 230000005540 biological transmission Effects 0.000 description 1
- 230000000295 complement effect Effects 0.000 description 1
- 230000006835 compression Effects 0.000 description 1
- 238000007906 compression Methods 0.000 description 1
- 210000004709 eyebrow Anatomy 0.000 description 1
- 210000000887 face Anatomy 0.000 description 1
- 239000011521 glass Substances 0.000 description 1
- 238000003384 imaging method Methods 0.000 description 1
- 230000006872 improvement Effects 0.000 description 1
- 230000010354 integration Effects 0.000 description 1
- 238000004519 manufacturing process Methods 0.000 description 1
- 238000013507 mapping Methods 0.000 description 1
- 238000005070 sampling Methods 0.000 description 1
- 238000011895 specific detection Methods 0.000 description 1
- 230000009466 transformation Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/40—Scenes; Scene-specific elements in video content
- G06V20/41—Higher-level, semantic clustering, classification or understanding of video scenes, e.g. detection, labelling or Markovian modelling of sport events or news items
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T1/00—General purpose image data processing
- G06T1/0021—Image watermarking
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/40—Scenes; Scene-specific elements in video content
- G06V20/46—Extracting features or characteristics from the video content, e.g. video fingerprints, representative shots or key frames
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2201/00—General purpose image data processing
- G06T2201/005—Image watermarking
- G06T2201/0065—Extraction of an embedded watermark; Reliable detection
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Multimedia (AREA)
- Computational Linguistics (AREA)
- Software Systems (AREA)
- Image Analysis (AREA)
Abstract
The embodiment of the application provides a video detection method, video detection equipment and a storage medium, wherein the method comprises the following steps: acquiring at least one frame of image in a video to be detected; extracting steganographic features and visual features associated with a specified object in at least one frame of image; analyzing the characteristic difference of at least one frame of image in time sequence according to the steganographic characteristics in at least one frame of image and the visual characteristics related to the specified object; and detecting whether the video to be detected is changed or not according to the steganographic features in the at least one frame of image, the visual features related to the specified object and the feature difference of the at least one frame of image in time sequence. In the embodiment, the detection of the spatial dimension and the time sequence dimension can be performed on the video to be detected, so that whether the video to be detected is changed or not can be determined by combining the detection conditions of the spatial dimension and the time dimension, and the comprehensiveness and the accuracy of video detection can be effectively improved.
Description
Technical Field
The present application relates to the field of video processing technologies, and in particular, to a video detection method, a video detection device, and a storage medium.
Background
At present, the video counterfeiting technology can forge the behavior expression of characters in the video at will by means of replacing faces in the video, changing expressions and the type of a grafting interface and the like.
Video counterfeiting causes serious interference to the fields of judicial investigation, insurance authentication, network security and the like. In these fields, the authenticity of video is particularly important, and therefore, detecting the authenticity of video becomes an urgent problem to be solved.
Disclosure of Invention
Aspects of the present disclosure provide a video detection method, apparatus, and storage medium to detect whether a video is altered.
The embodiment of the application provides a video detection method, which comprises the following steps:
acquiring at least one frame of image in a video to be detected;
extracting steganographic features and visual features associated with a specified object in the at least one frame of image;
analyzing the time-sequence feature difference of the at least one frame of image according to the steganographic features in the at least one frame of image and the visual features associated with the specified objects;
and detecting whether the video to be detected is changed or not according to the steganographic features in the at least one frame of image, the visual features related to the specified object and the feature difference of the at least one frame of image in time sequence.
The embodiment of the application also provides a computing device, which comprises a memory and a processor;
the memory is to store one or more computer instructions;
the processor is coupled with the memory for executing the one or more computer instructions for:
acquiring at least one frame of image in a video to be detected;
extracting steganographic features and visual features associated with a specified object in the at least one frame of image;
analyzing the time-sequence feature difference of the at least one frame of image according to the steganographic features in the at least one frame of image and the visual features associated with the specified objects;
and detecting whether the video to be detected is changed or not according to the steganographic features in the at least one frame of image, the visual features related to the specified object and the feature difference of the at least one frame of image in time sequence.
Embodiments of the present application also provide a computer-readable storage medium storing computer instructions that, when executed by one or more processors, cause the one or more processors to perform the aforementioned video detection method.
In the embodiment of the application, in the process of detecting a video to be detected, the steganographic features in the single-frame image and the visual features associated with the designated object can be extracted, and the feature difference between the frame images in time sequence can be analyzed according to the steganographic features in the single-frame image and the visual features associated with the designated object, so that whether the designated object in the video to be detected is changed or not can be determined according to the steganographic features in the single-frame image, the visual features associated with the designated object and the feature difference between the frame images in time sequence. Accordingly, in this embodiment, the video to be detected can be detected in the spatial dimension based on the steganographic features in the single-frame image and the visual features associated with the designated object, and the video to be detected can be detected in the time sequence dimension based on the feature difference between the frame images in the time sequence, so that whether the video to be detected is changed or not can be determined by combining the detection conditions of the spatial dimension and the time dimension, which can effectively improve the comprehensiveness and accuracy of video detection.
Drawings
The accompanying drawings, which are included to provide a further understanding of the application and are incorporated in and constitute a part of this application, illustrate embodiment(s) of the application and together with the description serve to explain the application and not to limit the application. In the drawings:
fig. 1a is a schematic flowchart of a video detection method according to an embodiment of the present application;
fig. 1b is a schematic view of a video detection process according to an embodiment of the present application;
fig. 2 is a schematic diagram of an application scenario provided in an embodiment of the present application;
fig. 3 is a schematic structural diagram of a computing device according to another embodiment of the present application.
Detailed Description
In order to make the objects, technical solutions and advantages of the present application more apparent, the technical solutions of the present application will be described in detail and completely with reference to the following specific embodiments of the present application and the accompanying drawings. It should be apparent that the described embodiments are only some of the embodiments of the present application, and not all of the embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present application.
At present, the video detection technology is not complete, and whether the video is counterfeit or not can not be effectively detected. In order to solve the problems existing in the prior art, in some embodiments of the present application: the video to be detected can be detected in the spatial dimension based on the steganographic features in the single-frame images and the visual features associated with the designated objects, and the video to be detected can be detected in the time dimension based on the feature difference between the frame images in the time sequence, so that whether the video to be detected is changed or not can be determined by combining the detection conditions of the spatial dimension and the time dimension, and the comprehensiveness and the accuracy of video detection can be effectively improved.
The technical solutions provided by the embodiments of the present application are described in detail below with reference to the accompanying drawings.
Fig. 1a is a schematic flowchart of a video detection method according to an embodiment of the present disclosure. As shown in fig. 1a, the method comprises:
100. acquiring at least one frame of image in a video to be detected;
101. extracting steganographic features and visual features associated with a specified object in at least one frame of image;
102. analyzing the characteristic difference of at least one frame of image in time sequence according to the steganographic characteristics in at least one frame of image and the visual characteristics related to the specified object;
103. and detecting whether the video to be detected is changed or not according to the steganographic features in the at least one frame of image, the visual features related to the specified object and the feature difference of the at least one frame of image in time sequence.
The video detection method provided by the embodiment can be applied to various scenes needing video detection, such as judicial investigation, insurance authentication, network security and the like. In various application scenarios, the designated objects may be set as desired. In this embodiment, the designated object may be a human face, a license plate number, a trademark, or other objects that may be changed, and is not limited herein.
Fig. 1b is a schematic view of a video detection process according to an embodiment of the present application, where fig. 1b corresponds to fig. 1 a. In fig. 1a and 1b, at least one frame of image in a video to be detected can be acquired. The at least one frame image may be a group of continuous frame images in the video to be detected, or a group of frame images selected through sampling, or of course, a frame image in which the designated object extracted from the video to be detected is located, which is not limited in this embodiment. In addition, the video to be detected can be a finished product video or a video in editing.
Based on the words, the obtained at least one frame of image can be subjected to steganographic feature extraction and visual feature extraction respectively so as to obtain the steganographic features and the visual features associated with the specified object in the at least one frame of image.
The steganographic features refer to inherent characteristic information hidden in data in the digital image imaging process. The steganographic features may be neighborhood correlation, color filter array CFA, lens characteristics, JPEG compression features, or the like, which is not limited in this embodiment. The domain correlation can represent the relationship between a pixel point in the image and the surrounding pixel points.
The visual feature (Frame feature) is image attribute information of an area in which a specified object is located in an image. The visual features may be edge contrast, etc., which is not limited in this embodiment.
It should be noted that, as mentioned above, at least one frame of image may be a set of consecutive frame images in the video to be detected, and therefore, there may be a frame image that does not include the designated object in at least one frame of image, in this case, for the frame image that does not include the designated object, in order to simplify the operation, the operation of extracting the visual feature associated with the designated object may not be performed, and of course, this embodiment does not limit this, and the operation may also be performed.
Accordingly, steganographic features and visual features associated with the specified object in at least one frame of image can be extracted. For example, when the designated object is a human face, the steganographic feature is a neighborhood correlation, and the visual feature is an edge contrast, the neighborhood correlation and the edge contrast associated with the human face in at least one frame of image may be extracted based on step 101.
In this embodiment, the temporal feature difference of the at least one frame of image may be analyzed according to the steganographic feature in the at least one frame of image and the visual feature associated with the designated object based on the temporal relationship existing between the at least one frame of image.
Specifically, the temporal analysis may be performed on the steganographic features in each frame of image according to the steganographic features in at least one frame of image, and the temporal analysis may be performed on the visual features associated with the designated object in each frame of image according to the visual features associated with the designated object in at least one frame of image, so as to determine the feature difference of each frame of image in temporal.
As a result, the time sequence of the neighborhood correlation and the edge contrast associated with the human face in each frame of image can be analyzed to determine the feature difference of each frame of image in time sequence.
On the basis, whether the at least one frame of image is changed or not can be detected according to the steganographic features in the at least one frame of image, the visual features related to the specified object and the feature difference of the at least one frame of image in time sequence, and then whether the video to be detected is changed or not can be detected.
When the steganographic features in at least one frame of image are abnormal, at least one frame of image can be changed.
When there is an abnormality in the visual feature associated with the specified object in the at least one image, the at least one image may be altered.
When the difference of the characteristics of the at least one frame of image in the time sequence is abnormal, the at least one frame of image may be altered.
Accordingly, in the embodiment, whether the at least one frame of image is changed or not can be detected by using the steganographic feature in the at least one frame of image, the visual feature associated with the specified object, and the feature difference of the at least one frame of image in time sequence, so as to determine whether the video to be detected is changed or not.
The change of the video to be detected means that the video to be detected is subjected to face replacement, license plate replacement, face filter, trademark replacement, license plate covering, character replacement and the like. Of course, the present embodiment is not limited thereto.
In addition, it should be noted that, in this embodiment, at least one frame of image may be used as a detection unit for video detection. It should be understood that, in this embodiment, a plurality of sets of frame images may be obtained from the image to be detected, and whether each set of frame image is changed or not may be detected, so as to determine whether the video to be detected is changed or not. Of course, the embodiment is not limited thereto, and enough frame images may be obtained from the image to be detected, and whether the video to be detected is changed or not may be determined directly according to the detection result of the frame images.
In this embodiment, in the process of detecting a video to be detected, the steganographic features and the visual features associated with the designated object in the single-frame image may be extracted, and the feature difference between the frame images in the time sequence may be analyzed according to the steganographic features and the visual features associated with the designated object in the single-frame image, so that whether the designated object in the video to be detected is modified or not may be determined according to the steganographic features in the single-frame image, the visual features associated with the designated object, and the feature difference between the frame images in the time sequence. Accordingly, in this embodiment, the video to be detected can be detected in the spatial dimension based on the steganographic features in the single-frame image and the visual features associated with the designated object, and the video to be detected can be detected in the time sequence dimension based on the feature difference between the frame images in the time sequence, so that whether the video to be detected is changed or not can be determined by combining the detection conditions of the spatial dimension and the time dimension, which can effectively improve the comprehensiveness and accuracy of video detection.
In the above or below embodiments, the at least one frame of image may be input to a convolutional neural network, and the steganographic features and the visual features associated with the specified object in the at least one frame of image may be extracted using the convolutional neural network.
In this embodiment, a convolutional neural network is used to perform extraction processing of steganographic features and visual features associated with a specified object.
In practical application, at least one frame of image can be input into a first convolution neural network, and the first convolution neural network is utilized to carry out convolution processing on the at least one frame of image so as to extract a feature map reflecting steganographic features from the at least one frame of image; and inputting the at least one frame of image into a second convolution neural network, and performing convolution processing on the at least one frame of image by using the second convolution neural network so as to extract a feature map reflecting visual features associated with the specified object from the at least one frame of image.
In this embodiment, a greater or lesser number of convolutional neural networks may also be used to perform the extraction processing of the steganographic features and the visual features associated with the designated object, which is not limited in this embodiment.
As mentioned above, the present embodiment does not limit what kind of steganographic features and visual features are used. In this regard, in this embodiment, different first convolutional neural networks may be used for different steganographic features, and different second convolutional neural networks may be used for different visual features.
Taking the case that the steganographic features adopt neighborhood correlation as an example, for each frame of image, performing convolution processing on the frame of image in sequence by utilizing a plurality of convolution layers of a first convolution neural network so as to extract a feature map reflecting the neighborhood correlation from the frame of image; in the first convolution layer of the first convolution neural network, convolution processing is carried out on the frame image by using a convolution kernel with the average value of 0.
In this example, the convolution kernels in the first convolutional layer of the first convolutional neural network are constrained such that the mean of each convolution kernel in that convolutional layer is 0. Based on this, when the frame image is convolved in the convolution layer, if the value change in the receptive field of the convolution kernel is gentle, that is, the value is close, the output value after convolution will be close to 0; if the value change in the field of the convolution kernel is more drastic, that is, the value difference is larger, the larger the convolved output value will be. Accordingly, the convolutional layer can extract the noise condition in the frame image, so that the first convolutional layer can limit the subsequent convolution processing process to the domain of the neighborhood correlation.
The subsequent convolutional layer of the first convolutional neural network can be utilized to continuously carry out convolution processing on the output of the first convolutional layer so as to carry out combination abstraction on the shallow layer features extracted by the first convolutional layer, and therefore the neighborhood correlation of the frame image is obtained.
Accordingly, in this example, the constraints may be designed in a data-driven fashion such that the first convolutional neural network may adaptively extract neighborhood correlations in at least one frame of the image.
Taking the visual feature as an example of the edge contrast, the frame image may be sequentially convolved by the plurality of convolution layers of the second convolutional neural network for each frame image, so as to extract a feature map reflecting the edge contrast associated with the specified object from the frame image. The specific convolution process is not described in detail herein.
It is worth noting that in this embodiment, multiple steganographic features and multiple visual features in at least one frame of image can be extracted simultaneously, different first convolutional neural networks can be configured for different steganographic features, and different second convolutional neural networks can be configured for different visual features.
Accordingly, in the embodiment, the constrained first convolutional neural network can be used to extract local and bottom-layer noise in the frame image, so that steganographic features in the frame image can be obtained adaptively; a second convolutional neural network may also be utilized to extract visual features associated with the specified object in the frame image. The steganographic features and the visual features are complementary, so that alteration traces in the frame image can be more comprehensively found based on the steganographic features in the frame image and the visual features associated with the specified object.
In the above or below embodiments, the steganographic features and the visual features associated with the designated object in the at least one frame of image may be respectively subjected to feature integration within the frame of image, so as to analyze the feature difference of the at least one frame of image in time sequence based on the integrated features.
In practical application, spatial feature vectors corresponding to at least one frame of image can be respectively generated according to steganographic features in at least one frame of image and visual features associated with a specified object; sequencing the space characteristic vectors corresponding to at least one frame of image according to the time sequence relation among at least one frame of image to form a space characteristic sequence; and analyzing the characteristic difference of at least one frame of image in time sequence according to the spatial characteristic sequence.
Wherein the spatial feature vector is used for representing the change trace in the frame image as a whole.
In this embodiment, the spatial feature sequence may be input into a recurrent neural network, and the cyclic neural network is utilized to analyze the feature difference of at least one frame of image in time sequence. When the feature difference of the at least one frame of image in time sequence is significant, the at least one frame of image may be altered.
For example, in a case that a certain frame of image in at least one frame of image is modified, the feature difference between the frame of image and the previous and subsequent frame of image will be significantly higher than the feature difference between the other frame of image, which is only one aspect, and the feature difference between the frame of image and the previous and subsequent frame of image may also be significantly higher than a certain preset value. Therefore, in various aspects, the recurrent neural network is used to analyze the feature difference of at least one frame of image in time sequence.
In this embodiment, multiple implementation manners may be adopted to generate the spatial feature vectors corresponding to the at least one frame of image.
In an optional implementation manner, for each frame image, global pooling processing can be performed on the steganographic features extracted from the frame image and the visual features associated with the designated object respectively to obtain steganographic feature vectors and visual feature vectors corresponding to the frame image; splicing the steganographic feature vector and the visual feature vector corresponding to the frame image, and inputting a splicing result into a first full-connection network; in the first full-connection network, the steganographic feature vector and the visual feature vector corresponding to the frame image are fused based on preset weight parameters to obtain a spatial feature vector corresponding to the frame image.
On the basis of extracting the steganographic features and the visual features by using the convolutional neural network mentioned in the foregoing embodiment, in this implementation, the feature map output by the first convolutional neural network and reflecting the steganographic features and the feature map output by the second convolutional neural network and reflecting the visual features associated with the specified object may be subjected to global pooling processing to obtain steganographic feature vectors and visual feature vectors corresponding to the frame image.
If the steganographic feature vector corresponding to the frame image is [ a ] and the visual feature vector is [ b ], the steganographic feature vector and the visual feature vector corresponding to the frame image can be spliced in a mode of [ a ], [ b ] } a, b ].
And presetting a weight parameter in the first fully-connected network, wherein the weight parameter is used for limiting the attention degree of the steganographic feature and the visual feature in the video detection process. And after the splicing result is input into a first full-connection network, the steganographic feature vector and the visual feature vector corresponding to the frame image can be fused by using the first full-connection network, so that the spatial feature vector corresponding to the frame image is obtained.
Of course, the above implementation is only exemplary, and the embodiment is not limited thereto.
In this embodiment, a scheme for fusing a steganographic feature and a visual feature is provided. In the spatial dimension, the steganographic features and the visual features are respectively extracted by different convolutional neural networks and then are fused into final spatial features through global pooling, feature splicing and full-connection networks. In the time dimension, different frame images can share a recurrent neural network, the fused spatial feature sequence is used as the input of the recurrent neural network, and the recurrent neural network can be used for analyzing additional time sequence features.
In the above or below embodiment, the steganographic features in the at least one frame of image and the difference in temporal characteristics between the visual features associated with the specified object and the at least one frame of image may be input into the second fully-connected network; and detecting the probability of the video to be detected being changed by utilizing the second full-connection network so as to determine whether the video to be detected is changed.
And the second full-connection network is preset with the mapping relation between the input information and the classification result. In this embodiment, the input information includes steganographic features in at least one frame of image and differences in visual features associated with the designated object and features of at least one frame of image in time sequence; the classification result may contain modified and unmodified.
Accordingly, the second fully-connected network may be utilized to predict a probability that at least one frame of the image is altered.
As mentioned above, in this embodiment, a plurality of groups of frame images in the network to be detected can be detected, and accordingly, whether the video to be detected is modified can be determined according to the modified probability corresponding to each group of frame images. For example, whether the video to be detected is modified can be determined according to the average value, the maximum value, and the like of the modified probabilities corresponding to the groups of frame images. And when the average value or the maximum value meets the changed condition, determining that the video to be detected is changed.
In the foregoing embodiment, on the basis of analyzing the feature difference of at least one frame of image in time sequence by using the recurrent neural network, in this embodiment, all output nodes of the recurrent neural network may be connected to the second fully-connected network, and the output of the recurrent neural network at each time may be provided to the second fully-connected network; in addition, based on the memory characteristics of the recurrent neural network, the recurrent neural network can not only input the characteristic difference of the at least one frame of image in time sequence into the second fully-connected network, but also provide the steganographic characteristics and the visual characteristics associated with the specified object in the at least one frame of image memorized by the recurrent neural network into the second fully-connected network.
Of course, this connection structure is merely exemplary, and in this embodiment, the convolutional neural network may also be connected to a second fully-connected network to provide the steganographic features and the visual features associated with the designated object in the at least one frame of image to the second fully-connected network. This embodiment is not limited to this.
In the above or below embodiments, the specified object in at least one frame of image may also be detected; carrying out alignment processing on a specified object in at least one frame of image; and inputting the at least one frame of image subjected to the alignment processing into a convolutional neural network so as to extract the steganographic features and the visual features associated with the specified object in the at least one frame of image by using the convolutional neural network.
Taking the designated object as the face feature as an example, in this embodiment, a face detection technology may be used to detect a face in at least one frame of image, and perform an alignment operation on the face.
Specifically, key feature points of the human face, such as glasses, nose tips, mouth corner points, eyebrows, contour points of each part of the human face, and the like, can be located. And aligning the human face in at least one frame of image through affine transformation based on the located key feature points.
Therefore, even if the position of the face deviates in different frame images, based on the alignment operation, the convolutional neural network does not need to pay attention to the deviation of the position of the face in different frame images as a special feature, so that the influence of the feature on the detection result can be avoided.
In the above or following embodiments, the video detection scheme may be implemented by using a video detection system, where the video detection system may include the convolutional neural network, the cyclic neural network, the fully-connected layer, and the like mentioned in the above embodiments, outputs of the first convolutional neural network and the second convolutional neural network are connected to the first fully-connected network, an output of the first fully-connected network is connected to the cyclic neural network, and an output of the cyclic neural network is connected to the second fully-connected network.
Based on the above framework of the video detection system, in this embodiment, the video detection system can be trained to determine parameters in the video detection system.
In practical application, an end-to-end mode can be adopted to carry out overall training on the video detection system.
Inputting training samples into a first convolution neural network and a second convolution neural network, wherein the first convolution neural network extracts the steganography characteristic of each frame of image, the second convolution neural network extracts the visual characteristic of each frame of image, the first full-connection network respectively fuses the steganography characteristic and the visual characteristic of each frame of image, the fused characteristics of all frames are input into a circulation neural network according to a time sequence, the output of the circulation neural network is subjected to time sequence average pooling, and the merged characteristics are connected to the second full-connection network.
And measuring the output result of the second fully-connected network by using a cross entropy loss function, and updating the parameters of the whole network. Wherein, the training sample is at least one frame image which is known whether to be changed or not.
Of course, the above training process is only exemplary, and the embodiment is not limited thereto.
In the above or following embodiments, it may also be determined whether the video to be detected carries pre-buried feature prompt information. In practical application, the appointed field in the video to be detected can be read, and whether the appointed field contains the pre-buried characteristic prompt information or not is judged.
And if the pre-buried characteristic prompt information is carried in the video to be detected, extracting the pre-buried characteristic of at least one frame of image.
The pre-buried feature is a user-defined feature which is hidden and configured in the video to be detected. For example, the pre-buried feature may be a watermark feature, a digital signature feature, or a white noise feature, among others.
In addition, in this embodiment, a convolutional neural network may be used to extract the pre-buried features, which is not limited to this embodiment.
Based on the method, whether the video to be detected is changed or not can be detected according to the pre-buried feature extraction result.
In this embodiment, the pre-buried feature extraction result at least includes that the pre-buried feature is not extracted, that the extracted pre-buried feature is abnormal is determined, or that the extracted pre-buried feature is not abnormal. In practical application, the extracted pre-buried features may be subjected to tamper detection to determine whether the extracted pre-buried features are abnormal, which is not limited in this embodiment.
Accordingly, in this embodiment, whether the video to be detected is modified can be detected according to the steganographic features in the at least one frame of image, the visual features associated with the designated object, the feature difference of the at least one frame of image in the time sequence, and the pre-buried feature extraction result.
The specific detection rule can be set according to actual needs.
For example, the probability that the video to be detected is changed can be predicted according to the steganographic features in at least one frame of image, the visual features related to the specified object and the feature difference of at least one frame of image in time sequence, and if the pre-buried features are not extracted or the extracted pre-buried features are determined to be abnormal in the pre-buried feature extraction result, the probability can be linearly compensated to improve the probability.
For another example, it may be determined that the video to be detected is changed when any one of the probability or the pre-buried feature extraction result does not meet the preset standard. For example, when the probability is higher than the preset standard of 0.5 or the pre-buried features are not extracted, it is determined that the video to be detected is changed.
For another example, it may be determined that the video to be detected is changed under the condition that the probability or the pre-buried feature extraction result does not meet the preset standard.
The above are exemplary, and the present embodiment is not limited thereto.
Fig. 2 is a schematic view of an application scenario of video detection according to an embodiment of the present application. As shown in fig. 2, the video detection scheme provided by the present embodiment is used to detect whether a face in a video is modified.
In the application scenario shown in fig. 2, the input is a video to be detected, and the output is whether the video to be detected is altered.
The following describes the video detection process with reference to fig. 2 by taking 5 consecutive frame images in the video to be detected as an example.
After 5 continuous frame images are preprocessed through face detection, extraction, alignment and the like, the preprocessed images are respectively input into CNN1 and CNN2, the CNN1 is used for extracting feature maps which respectively correspond to the 5 continuous frame images and reflect neighborhood correlation, and the CNN2 is used for extracting edge contrast which respectively corresponds to the 5 continuous frame images and reflects the correlation with the face.
The output of CNN1 and CNN2 is input into the first fully-connected network after global pooling and feature splicing. And fusing neighborhood correlation and edge contrast associated with the human face in 5 continuous frame images into 5 spatial feature vectors by using a first full-connection network, and sequencing the 5 spatial feature vectors based on the time sequence relation of the 5 continuous frame images to form a spatial feature sequence.
And inputting the spatial feature sequence into the RNN to analyze the feature difference of 5 continuous frame images in time sequence by using the RNN.
And the RNN inputs the analyzed feature difference of the 5 continuous frame images on the time sequence, the memorized neighborhood correlation and the edge contrast related to the human face in the 5 continuous frame images into the second full-connection layer.
The second fully-connected layer may output the probability that 5 consecutive frame images are altered.
After the video detection process is carried out on other frame images in the video to be detected, whether the video to be detected is changed or not can be determined. For example, whether the video to be detected is altered can be defined by setting a probability threshold.
It should be noted that the execution subjects of the steps of the methods provided in the above embodiments may be the same device, or different devices may be used as the execution subjects of the methods. For example, the execution subjects of steps 101 to 102 may be device a; for another example, the execution subject of steps 101 and 102 may be device a, and the execution subject of step 103 may be device B; and so on.
In addition, in some of the flows described in the above embodiments and the drawings, a plurality of operations are included in a specific order, but it should be clearly understood that the operations may be executed out of the order presented herein or in parallel, and the sequence numbers of the operations, such as 101, 102, etc., are merely used for distinguishing different operations, and the sequence numbers do not represent any execution order per se. Additionally, the flows may include more or fewer operations, and the operations may be performed sequentially or in parallel. It should be noted that, the descriptions of "first", "second", etc. in this document are used for distinguishing different messages, devices, modules, etc., and do not represent a sequential order, nor limit the types of "first" and "second" to be different.
Fig. 3 is a schematic structural diagram of a computing device according to another embodiment of the present application. As shown in fig. 3, the computing device includes: a memory 30 and a processor 31.
The memory 30 is used to store computer programs and may be configured to store other various data to support operations on the computing device. Examples of such data include instructions for any application or method operating on the computing device, contact data, phonebook data, messages, pictures, videos, and so forth.
The memory 30 may be implemented by any type or combination of volatile or non-volatile memory devices, such as Static Random Access Memory (SRAM), electrically erasable programmable read-only memory (EEPROM), erasable programmable read-only memory (EPROM), programmable read-only memory (PROM), read-only memory (ROM), magnetic memory, flash memory, magnetic or optical disks.
A processor 31, coupled to the memory 30, for executing the computer program in the memory 30 for:
acquiring at least one frame of image in a video to be detected;
extracting steganographic features and visual features associated with a specified object in at least one frame of image;
analyzing the characteristic difference of at least one frame of image in time sequence according to the steganographic characteristics in at least one frame of image and the visual characteristics related to the specified object;
and detecting whether the video to be detected is changed or not according to the steganographic features in the at least one frame of image, the visual features related to the specified object and the feature difference of the at least one frame of image in time sequence.
In an alternative embodiment, the processor 31, when extracting the steganographic features and the visual features associated with the designated object in at least one frame of the image, is configured to:
and inputting the at least one frame of image into a convolutional neural network, and extracting the steganographic features and the visual features associated with the specified object in the at least one frame of image by using the convolutional neural network.
In an alternative embodiment, the processor 31, when inputting the at least one frame of image into the convolutional neural network, and extracting the steganographic features and the visual features associated with the designated object in the at least one frame of image using the convolutional neural network, is configured to:
inputting at least one frame of image into a first convolution neural network, and performing convolution processing on the at least one frame of image by using the first convolution neural network so as to extract a feature map reflecting steganographic features from the at least one frame of image;
and inputting the at least one frame of image into a second convolution neural network, and performing convolution processing on the at least one frame of image by using the second convolution neural network so as to extract a feature map reflecting visual features associated with the specified object from the at least one frame of image.
In an alternative embodiment, the steganographic feature is a neighborhood correlation, and the processor 31, when performing convolution processing on the at least one frame of image by using the first convolution neural network to extract a feature map reflecting the steganographic feature from the at least one frame of image, is configured to:
and sequentially performing convolution processing on each frame image by utilizing a plurality of convolution layers of the first convolution neural network so as to extract a characteristic diagram reflecting neighborhood correlation from the frame image, wherein in the first convolution layer of the first convolution neural network, convolution kernel with the average value of 0 is used for performing convolution processing on the frame image.
In an alternative embodiment, the processor 31, when inputting the at least one frame of image into the convolutional neural network, is configured to:
detecting a designated object in at least one frame of image;
carrying out alignment processing on a specified object in at least one frame of image;
and inputting the at least one frame of image subjected to the alignment processing into a convolutional neural network so as to extract the steganographic features and the visual features associated with the specified object in the at least one frame of image by using the convolutional neural network.
In an alternative embodiment, the processor 31 is configured to, when analyzing feature variability of at least one frame of image in time sequence according to the steganographic features in the at least one frame of image and the visual features associated with the designated object:
respectively generating spatial feature vectors corresponding to the at least one frame of image according to the steganographic features in the at least one frame of image and the visual features associated with the designated object;
sequencing the space characteristic vectors corresponding to at least one frame of image according to the time sequence relation among at least one frame of image to form a space characteristic sequence;
and analyzing the characteristic difference of at least one frame of image in time sequence according to the spatial characteristic sequence.
In an optional embodiment, the processor 31, when generating the spatial feature vectors corresponding to the at least one frame of image respectively according to the steganographic features in the at least one frame of image and the visual features associated with the designated object, is configured to:
aiming at each frame of image, global pooling processing is respectively carried out on the steganographic features extracted from the frame image and the visual features associated with the designated object so as to obtain steganographic feature vectors and visual feature vectors corresponding to the frame image;
splicing the steganographic feature vector and the visual feature vector corresponding to the frame image, and inputting a splicing result into a first full-connection network;
in the first full-connection network, the steganographic feature vector and the visual feature vector corresponding to the frame image are fused based on preset weight parameters to obtain a spatial feature vector corresponding to the frame image.
In an alternative embodiment, the processor 31, when analyzing the temporal feature difference of at least one frame of image according to the spatial feature sequence, is configured to:
and inputting the spatial feature sequence into a recurrent neural network, and analyzing the feature difference of at least one frame of image in time sequence by using the recurrent neural network.
In an optional embodiment, the processor 31, when detecting whether the video to be detected is altered according to the steganographic features in the at least one frame of image and the difference between the visual features associated with the specified object and the features of the at least one frame of image in time sequence, is configured to:
inputting steganographic features in the at least one frame of image and visual features associated with the designated object and feature differences of the at least one frame of image in time sequence into a second fully-connected network;
and detecting the probability of the video to be detected being changed by utilizing the second full-connection network so as to determine whether the video to be detected is changed.
In an alternative embodiment, the designated object is a face, a license plate number, a text or a trademark.
In an alternative embodiment, the processor 31 is further configured to:
if the pre-buried characteristic prompt information is carried in the video to be detected, extracting pre-buried characteristics of the at least one frame of image, wherein the pre-buried characteristics are self-defined characteristics which are hidden and configured in the video to be detected;
and detecting whether the video to be detected is changed or not according to the pre-buried feature extraction result.
In an optional embodiment, the pre-buried feature extraction result is that the pre-buried feature is not extracted, the extracted pre-buried feature is determined to have abnormality, or the extracted pre-buried feature is determined to have no abnormality.
In an optional embodiment, the pre-buried feature is one or more of a watermark feature, a digital signature feature, or a white noise feature.
Further, as shown in fig. 3, the computing device further includes: communication components 32, power components 33, and the like. Only some of the components are schematically shown in fig. 3, and the computing device is not meant to include only the components shown in fig. 3.
Wherein the communication component 32 is configured to facilitate wired or wireless communication between the device in which the communication component is located and other devices. The device in which the communication component is located may access a wireless network based on a communication standard, such as WiFi, 2G or 3G, or a combination thereof. In an exemplary embodiment, the communication component receives a broadcast signal or broadcast related information from an external broadcast management system via a broadcast channel. In an exemplary embodiment, the communication component may be implemented based on Near Field Communication (NFC) technology, Radio Frequency Identification (RFID) technology, infrared data association (IrDA) technology, Ultra Wideband (UWB) technology, Bluetooth (BT) technology, or other technologies to facilitate short-range communications.
The power supply unit 33 supplies power to various components of the device in which the power supply unit is installed. The power components may include a power management system, one or more power supplies, and other components associated with generating, managing, and distributing power for the device in which the power component is located.
Accordingly, the present application further provides a computer-readable storage medium storing a computer program, where the computer program can implement the steps that can be executed by a computing device in the foregoing method embodiments when executed.
As will be appreciated by one skilled in the art, embodiments of the present invention may be provided as a method, system, or computer program product. Accordingly, the present invention may take the form of an entirely hardware embodiment, an entirely software embodiment or an embodiment combining software and hardware aspects. Furthermore, the present invention may take the form of a computer program product embodied on one or more computer-usable storage media (including, but not limited to, disk storage, CD-ROM, optical storage, and the like) having computer-usable program code embodied therein.
The present invention is described with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems), and computer program products according to embodiments of the invention. It will be understood that each flow and/or block of the flow diagrams and/or block diagrams, and combinations of flows and/or blocks in the flow diagrams and/or block diagrams, can be implemented by computer program instructions. These computer program instructions may be provided to a processor of a general purpose computer, special purpose computer, embedded processor, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions specified in the flowchart flow or flows and/or block diagram block or blocks.
These computer program instructions may also be stored in a computer-readable memory that can direct a computer or other programmable data processing apparatus to function in a particular manner, such that the instructions stored in the computer-readable memory produce an article of manufacture including instruction means which implement the function specified in the flowchart flow or flows and/or block diagram block or blocks.
These computer program instructions may also be loaded onto a computer or other programmable data processing apparatus to cause a series of operational steps to be performed on the computer or other programmable apparatus to produce a computer implemented process such that the instructions which execute on the computer or other programmable apparatus provide steps for implementing the functions specified in the flowchart flow or flows and/or block diagram block or blocks.
In a typical configuration, a computing device includes one or more processors (CPUs), input/output interfaces, network interfaces, and memory.
The memory may include forms of volatile memory in a computer readable medium, Random Access Memory (RAM) and/or non-volatile memory, such as Read Only Memory (ROM) or flash memory (flash RAM). Memory is an example of a computer-readable medium.
Computer-readable media, including both non-transitory and non-transitory, removable and non-removable media, may implement information storage by any method or technology. The information may be computer readable instructions, data structures, modules of a program, or other data. Examples of computer storage media include, but are not limited to, phase change memory (PRAM), Static Random Access Memory (SRAM), Dynamic Random Access Memory (DRAM), other types of Random Access Memory (RAM), Read Only Memory (ROM), Electrically Erasable Programmable Read Only Memory (EEPROM), flash memory or other memory technology, compact disc read only memory (CD-ROM), Digital Versatile Discs (DVD) or other optical storage, magnetic cassettes, magnetic tape magnetic disk storage or other magnetic storage devices, or any other non-transmission medium that can be used to store information that can be accessed by a computing device. As defined herein, a computer readable medium does not include a transitory computer readable medium such as a modulated data signal and a carrier wave.
It should also be noted that the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Without further limitation, an element defined by the phrase "comprising an … …" does not exclude the presence of other like elements in a process, method, article, or apparatus that comprises the element.
The above description is only an example of the present application and is not intended to limit the present application. Various modifications and changes may occur to those skilled in the art. Any modification, equivalent replacement, improvement, etc. made within the spirit and principle of the present application should be included in the scope of the claims of the present application.
Claims (27)
1. A video detection method, comprising:
acquiring at least one frame of image in a video to be detected;
extracting steganographic features and visual features associated with a specified object in the at least one frame of image;
analyzing the time-sequence feature difference of the at least one frame of image according to the steganographic features in the at least one frame of image and the visual features associated with the specified objects;
and detecting whether the video to be detected is changed or not according to the steganographic features in the at least one frame of image, the visual features related to the specified object and the feature difference of the at least one frame of image in time sequence.
2. The method of claim 1, wherein extracting the steganographic features and the visual features associated with the designated object in the at least one frame of image comprises:
and inputting the at least one frame of image into a convolutional neural network, and extracting the steganographic features and the visual features associated with the specified object in the at least one frame of image by using the convolutional neural network.
3. The method of claim 2, wherein inputting the at least one frame of image into a convolutional neural network, extracting steganographic features and visual features associated with a specified object in the at least one frame of image using the convolutional neural network, comprises:
inputting the at least one frame of image into a first convolution neural network, and performing convolution processing on the at least one frame of image by using the first convolution neural network so as to extract a feature map reflecting steganographic features from the at least one frame of image;
and inputting the at least one frame of image into a second convolutional neural network, and performing convolution processing on the at least one frame of image by using the second convolutional neural network so as to extract a feature map reflecting visual features associated with a specified object from the at least one frame of image.
4. The method of claim 3, wherein the steganographic features are neighborhood correlations, and wherein the convolving the at least one frame of image with the first convolutional neural network to extract a feature map reflecting the steganographic features from the at least one frame of image comprises:
for each frame image, sequentially performing convolution processing on the frame image by utilizing a plurality of convolution layers of the first convolution neural network so as to extract a feature map reflecting neighborhood correlation from the frame image;
and performing convolution processing on the frame image by using a convolution kernel with the mean value of 0 in a first convolution layer of the first convolution neural network.
5. The method of claim 2, wherein said inputting said at least one frame of image into a convolutional neural network comprises:
detecting a designated object in the at least one frame of image;
carrying out alignment processing on a specified object in the at least one frame of image;
inputting the at least one frame of image subjected to the alignment processing into the convolutional neural network so as to extract the steganographic features and the visual features associated with the specified object in the at least one frame of image by using the convolutional neural network.
6. The method according to claim 1, wherein analyzing the temporal feature variability of the at least one frame of image according to the steganographic features in the at least one frame of image and the visual features associated with the designated objects comprises:
respectively generating spatial feature vectors corresponding to the at least one frame of image according to the steganographic features in the at least one frame of image and the visual features associated with the designated object;
sequencing the space characteristic vectors corresponding to the at least one frame of image according to the time sequence relation among the at least one frame of image to form a space characteristic sequence;
and analyzing the characteristic difference of the at least one frame of image in time sequence according to the spatial characteristic sequence.
7. The method according to claim 6, wherein the generating spatial feature vectors corresponding to the at least one frame of image respectively according to the steganographic features in the at least one frame of image and the visual features associated with the designated object comprises:
aiming at each frame image, performing global pooling on the steganographic features extracted from the frame image and the visual features associated with the specified object respectively to obtain steganographic feature vectors and visual feature vectors corresponding to the frame image;
splicing the steganographic feature vector and the visual feature vector corresponding to the frame image, and inputting a splicing result into a first full-connection network;
and in the first full-connection network, fusing the steganographic feature vector and the visual feature vector corresponding to the frame image based on a preset weight parameter to obtain a spatial feature vector corresponding to the frame image.
8. The method of claim 6, wherein analyzing the temporal feature variability of the at least one frame of image according to the spatial feature sequence comprises:
inputting the spatial feature sequence into a recurrent neural network, and analyzing feature difference of the at least one frame of image in time sequence by using the recurrent neural network.
9. The method according to any one of claims 1 to 8, wherein the detecting whether the video to be detected is altered according to the steganographic features in the at least one frame of image and the difference between the visual features associated with the designated object and the features of the at least one frame of image in time sequence comprises:
inputting steganographic features in the at least one frame of image and differences in the visual features associated with the designated object and the features of the at least one frame of image in time sequence into a second fully connected network;
and detecting the probability of the video to be detected being changed by utilizing the second full-connection network so as to determine whether the video to be detected is changed.
10. The method according to any one of claims 1 to 8, wherein the designated object is a face, a letter, a license plate number, or a trademark.
11. The method of claim 1, further comprising:
if the pre-buried characteristic prompt information is carried in the video to be detected, extracting pre-buried characteristics of the at least one frame of image, wherein the pre-buried characteristics are self-defined characteristics which are hidden and configured in the video to be detected;
and detecting whether the video to be detected is changed or not according to the pre-buried feature extraction result.
12. The method of claim 11, wherein the pre-buried feature extraction result comprises: and extracting no embedded features, determining that the extracted embedded features are abnormal or determining that the extracted embedded features are not abnormal.
13. The method of claim 11, wherein the pre-buried features are one or more of watermark features, digital signature features, or white noise features.
14. A computing device comprising a memory and a processor;
the memory is to store one or more computer instructions;
the processor is coupled with the memory for executing the one or more computer instructions for:
acquiring at least one frame of image in a video to be detected;
extracting steganographic features and visual features associated with a specified object in the at least one frame of image;
analyzing the time-sequence feature difference of the at least one frame of image according to the steganographic features in the at least one frame of image and the visual features associated with the specified objects;
and detecting whether the video to be detected is changed or not according to the steganographic features in the at least one frame of image, the visual features related to the specified object and the feature difference of the at least one frame of image in time sequence.
15. The device of claim 14, wherein the processor, in extracting the steganographic features and the visual features associated with the designated object in the at least one frame of image, is configured to:
and inputting the at least one frame of image into a convolutional neural network, and extracting the steganographic features and the visual features associated with the specified object in the at least one frame of image by using the convolutional neural network.
16. The apparatus of claim 15, wherein the processor, in inputting the at least one frame of image into a convolutional neural network, extracting steganographic features and visual features associated with a specified object in the at least one frame of image using the convolutional neural network, is configured to:
inputting the at least one frame of image into a first convolution neural network, and performing convolution processing on the at least one frame of image by using the first convolution neural network so as to extract a feature map reflecting steganographic features from the at least one frame of image;
and inputting the at least one frame of image into a second convolutional neural network, and performing convolution processing on the at least one frame of image by using the second convolutional neural network so as to extract a feature map reflecting visual features associated with a specified object from the at least one frame of image.
17. The apparatus of claim 16, wherein the steganographic feature is a neighborhood correlation, and wherein the processor, when convolving the at least one frame of image with the first convolution neural network to extract a feature map from the at least one frame of image that reflects the steganographic feature, is configured to:
and for each frame image, sequentially performing convolution processing on the frame image by using a plurality of convolution layers of the first convolution neural network so as to extract a feature map reflecting neighborhood correlation from the frame image, wherein in the first convolution layer of the first convolution neural network, convolution processing is performed on the frame image by using a convolution core with the mean value of 0.
18. The apparatus of claim 13, wherein the processor, when inputting the at least one frame of image into the convolutional neural network, is configured to:
detecting a designated object in the at least one frame of image;
carrying out alignment processing on a specified object in the at least one frame of image;
inputting the at least one frame of image subjected to the alignment processing into the convolutional neural network so as to extract the steganographic features and the visual features associated with the specified object in the at least one frame of image by using the convolutional neural network.
19. The apparatus of claim 14, wherein the processor, when analyzing the temporal feature variability of the at least one frame of images based on the steganographic features in the at least one frame of images and the visual features associated with the designated objects, is configured to:
respectively generating spatial feature vectors corresponding to the at least one frame of image according to the steganographic features in the at least one frame of image and the visual features associated with the designated object;
sequencing the space characteristic vectors corresponding to the at least one frame of image according to the time sequence relation among the at least one frame of image to form a space characteristic sequence;
and analyzing the characteristic difference of the at least one frame of image in time sequence according to the spatial characteristic sequence.
20. The apparatus of claim 19, wherein the processor, when generating spatial feature vectors corresponding to the at least one frame of image respectively according to the steganographic features in the at least one frame of image and the visual features associated with the designated object, is configured to:
aiming at each frame of image, respectively carrying out global pooling on the steganographic features extracted from the frame of image and the visual features associated with the specified object so as to obtain steganographic feature vectors and visual feature vectors corresponding to the frame of image;
splicing the steganographic feature vector and the visual feature vector corresponding to the frame image, and inputting a splicing result into a first full-connection network;
and in the first full-connection network, fusing the steganographic feature vector and the visual feature vector corresponding to the frame image based on a preset weight parameter to obtain a spatial feature vector corresponding to the frame image.
21. The apparatus of claim 19, wherein the processor, when analyzing the temporal feature variability of the at least one frame of image according to the spatial feature sequence, is configured to:
inputting the spatial feature sequence into a recurrent neural network, and analyzing feature difference of the at least one frame of image in time sequence by using the recurrent neural network.
22. The apparatus according to any one of claims 14-21, wherein the processor, when detecting whether the video to be detected is altered according to the steganographic features in the at least one frame of image and the difference in the visual features associated with the designated object and the features of the at least one frame of image in time sequence, is configured to:
inputting steganographic features in the at least one frame of image and differences in the visual features associated with the designated object and the features of the at least one frame of image in time sequence into a second fully connected network;
and detecting the probability of the video to be detected being changed by utilizing the second full-connection network so as to determine whether the video to be detected is changed.
23. The apparatus according to any one of claims 14 to 21, wherein the designated object is a face, a letter, a license plate number, or a trademark.
24. The device of claim 14, wherein the processor is further configured to:
if the pre-buried characteristic prompt information is carried in the video to be detected, extracting pre-buried characteristics of the at least one frame of image, wherein the pre-buried characteristics are self-defined characteristics which are hidden and configured in the video to be detected;
and detecting whether the video to be detected is changed or not according to the pre-buried feature extraction result.
25. The apparatus of claim 24, wherein the pre-buried feature extraction result is that no pre-buried features are extracted, that the extracted pre-buried features are determined to be abnormal, or that the extracted pre-buried features are determined to be not abnormal.
26. The apparatus of claim 24, wherein the pre-buried feature is one or more of a watermark feature, a digital signature feature, or a white noise feature.
27. A computer-readable storage medium storing computer instructions, which when executed by one or more processors, cause the one or more processors to perform the video detection method of any one of claims 1-13.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201911006641.7A CN112699707B (en) | 2019-10-22 | 2019-10-22 | Video detection method, device and storage medium |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201911006641.7A CN112699707B (en) | 2019-10-22 | 2019-10-22 | Video detection method, device and storage medium |
Publications (2)
Publication Number | Publication Date |
---|---|
CN112699707A true CN112699707A (en) | 2021-04-23 |
CN112699707B CN112699707B (en) | 2024-05-28 |
Family
ID=75504575
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201911006641.7A Active CN112699707B (en) | 2019-10-22 | 2019-10-22 | Video detection method, device and storage medium |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN112699707B (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114648478A (en) * | 2022-03-29 | 2022-06-21 | 北京小米移动软件有限公司 | Image processing method, device, chip, electronic equipment and storage medium |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20100091981A1 (en) * | 2008-04-14 | 2010-04-15 | Yun-Qing Shi | Steganalysis of Suspect Media |
CN104519361A (en) * | 2014-12-12 | 2015-04-15 | 天津大学 | Video steganography analysis method based on space-time domain local binary pattern |
CN104837011A (en) * | 2015-05-04 | 2015-08-12 | 中国科学院信息工程研究所 | Content self-adaptive video steganalysis method |
US20190005657A1 (en) * | 2017-06-30 | 2019-01-03 | Baidu Online Network Technology (Beijing) Co., Ltd . | Multiple targets-tracking method and apparatus, device and storage medium |
CN110312138A (en) * | 2019-01-04 | 2019-10-08 | 北京大学 | A kind of high embedding capacity video steganography method and system based on the modeling of time series error convolution |
-
2019
- 2019-10-22 CN CN201911006641.7A patent/CN112699707B/en active Active
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20100091981A1 (en) * | 2008-04-14 | 2010-04-15 | Yun-Qing Shi | Steganalysis of Suspect Media |
CN104519361A (en) * | 2014-12-12 | 2015-04-15 | 天津大学 | Video steganography analysis method based on space-time domain local binary pattern |
CN104837011A (en) * | 2015-05-04 | 2015-08-12 | 中国科学院信息工程研究所 | Content self-adaptive video steganalysis method |
US20190005657A1 (en) * | 2017-06-30 | 2019-01-03 | Baidu Online Network Technology (Beijing) Co., Ltd . | Multiple targets-tracking method and apparatus, device and storage medium |
CN110312138A (en) * | 2019-01-04 | 2019-10-08 | 北京大学 | A kind of high embedding capacity video steganography method and system based on the modeling of time series error convolution |
Non-Patent Citations (3)
Title |
---|
SAMANEH SHAFEE; BOSHRA RAJAEI: "A secure steganography algorithm using compressive sensing based on HVS feature", IEEE, 2 November 2017 (2017-11-02) * |
叶俊辰;黄潇洒;王士林;: "基于卷积神经网络的nsF5隐写检测研究", 通信技术, no. 03, 10 March 2019 (2019-03-10) * |
周洋;何永健;唐向宏;陆宇;蒋刚毅;: "融合双目多维感知特征的立体视频显著性检测", 中国图象图形学报, no. 03, 16 March 2017 (2017-03-16) * |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114648478A (en) * | 2022-03-29 | 2022-06-21 | 北京小米移动软件有限公司 | Image processing method, device, chip, electronic equipment and storage medium |
Also Published As
Publication number | Publication date |
---|---|
CN112699707B (en) | 2024-05-28 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US10936919B2 (en) | Method and apparatus for detecting human face | |
CN110348439B (en) | Method, computer readable medium and system for automatically identifying price tags | |
CN111738349B (en) | Detection effect evaluation method and device of target detection algorithm, storage medium and equipment | |
CN109116129B (en) | Terminal detection method, detection device, system and storage medium | |
CN111415336B (en) | Image tampering identification method, device, server and storage medium | |
CN112633255B (en) | Target detection method, device and equipment | |
CN112101200A (en) | Human face anti-recognition method, system, computer equipment and readable storage medium | |
WO2023123981A1 (en) | Video processing method and apparatus, computer device and storage medium | |
CN111241873A (en) | Image reproduction detection method, training method of model thereof, payment method and payment device | |
CN111881740A (en) | Face recognition method, face recognition device, electronic equipment and medium | |
CN113313098B (en) | Video processing method, device, system and storage medium | |
CN110956102A (en) | Bank counter monitoring method and device, computer equipment and storage medium | |
CN112699707B (en) | Video detection method, device and storage medium | |
CN112907206B (en) | Business auditing method, device and equipment based on video object identification | |
CN113255766A (en) | Image classification method, device, equipment and storage medium | |
US20220327849A1 (en) | Computer-Implemented Method For Optical Character Recognition | |
CN113518217A (en) | Object recognition method, device, server and medium | |
CN113516615A (en) | Sample generation method, system, equipment and storage medium | |
CN113837174A (en) | Target object identification method and device and computer equipment | |
Salguero-Cruz et al. | Proposal of a Comparative Framework for Face Super-Resolution Algorithms in Forensics | |
CN117114711B (en) | Data analysis and management method and platform for tobacco monopoly law enforcement based on artificial intelligence | |
CN113542866B (en) | Video processing method, device, equipment and computer readable storage medium | |
CN113763283B (en) | Detection method and device for defogging of image and intelligent device | |
US20240037915A1 (en) | Method and system for preprocessing optimization of streaming video data using machine learning | |
CN110781348B (en) | Video file analysis method |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |