1. Introduction
Steganography, another term for covert communication, is a technique for the hidden messages in objects that do not easily arouse suspicion and then sending them to the intended recipients, and it has received widespread attention in the field of information security in recent years [
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14]. The digital medium “steganography” was used by illegal organizations to covertly communicate and engage in activities that hazard national security. Steganalysis, the corresponding attack technique, is to extract the hidden messages in order to counter steganography and protect national security [
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26].
With digital media developing rapidly, improving the speed and accuracy of steganalysis has become a pressing problem. Therefore, the digital image adaptive steganalysis algorithm is the direction that researchers are currently focusing on, mainly by extracting the steganalysis features [
27,
28,
29,
30,
31,
32], and using the integrated classifier for training and detection, which can provide a detection effect. At present, researchers have developed a series of high-dimensional steganalysis algorithms. For example, Holub et al. [
27] proposed the 8000-D DCTR feature of low complexity (fast) extracted from the used DCT residuals, Song et al. [
28] proposed the 17,000-D GFR feature, in which the Gabor filter can provide steganalysis from different scales, Kodovský et al. [
29] proposed the 22,510-D CC-JRM feature established by a set of sub-model systems covering the joint distribution of spatial and frequency domain DCT coefficients with extensive statistical correlation, and Fridrich et al. [
30] proposed the 34,671-D SRM feature of the complete spatial domain rich model. Although high-dimensional steganalysis features achieve high detection accuracy for image-adaptive steganography, the adaptive steganalysis algorithm extracts steganalysis features with high dimension, which leads to high spatio-temporal complexity for detecting the stego images and affects the development of fast steganalysis. Therefore, how to select the feature components that contribute to detecting, so as to reduce the steganalysis feature dimension and the spatio-temporal complexity of detecting the stego images, has become the focus of the current steganalysis research.
At present, researchers have conducted a series of studies on the selection and dimension reduction of steganalysis features [
30,
33,
34,
35,
36,
37]. These methods can be classified into general and specific steganalysis feature selection methods depending on the objects to which the feature selection methods are applied. General methods are suitable for measuring a variety of steganalysis features, measuring the contribution of feature components to detection of the stego images, and selecting the feature vector with a large contribution to detecting of the stego images for training and detection. Typical methods in this type are the following. Qin et al. [
33] proposed a method with no parameter limit, which is simple to use, and in which it is easy to understand the results of the PCA (abbreviated as PCA-D method). This method calculates the covariance matrix and the feature values and feature vector of the covariance matrix, and finally determines the number of selected feature components by determining the number of principal components. However, this method has poor effect on the nonlinear structures data. Ma et al. [
34] proposed a feature selection method based on decision rough set
-positive domain simplification. The method not only reduces the feature dimension but maintains the detection accuracy of the stego images; however, the feature selection depends on the classifier results, which leads to a high time complexity of the selection. In our previous contribution [
35], we proposed a feature selection method (abbreviated as CGSM method) based on comprehensive criteria, which combine the difference function and correlation, and select the feature components with large difference and delete the feature components with weak correlation, which slightly improves the detection accuracy of the stego images and reduces the feature dimension.
Specific methods are suitable for a certain steganalysis feature. Though the calculation for this class of selection methods is simpler than those for the general class, it has a narrower scope of application. Typical methods of this kind are outlined here. Fridrich et al. [
30] proposed a series of SRM feature selection methods for airspace-rich models, and most of these take each sub-model of SRM as a subset and then select the features based on the diversity of classifiers. Among them, the BEST-q-CLASS method has the best detection performance in selecting features, reducing the feature dimension and improving the detection accuracy of SRM steganalysis features. However, this method ignores the redundancy between subsets when selecting features, so the feature dimension is still high after selection. Yang et al. [
36] proposed an algorithm for GFR feature subspace selection based on Fisher’s criterion (abbreviated as Fisher-G method). The method is able to select more efficient feature subspaces to improve the detection performance of GFR features on the stego images in a targeted way, however, it does not improve the detection accuracy of the stego images significantly when the quality factor is high. Yu et al. [
37] proposed a multi-scale GFR feature selection method based on the SNR criterion combined with an improved Relief algorithm (abbreviated as SRGS method), which deletes useless and low-contribution feature components and significantly reduces the feature dimension while maintaining the detection accuracy. However, it has no obvious effect on other steganalysis features.
So far, some studies have achieved different steganalysis feature selection effects [
28,
33,
34,
35,
36,
37], such as CC-PEV, GFR, CC-JRM, SRM and J+SRM features. However, the existing methods for DCTR feature selection effect are not satisfactory, the selected feature dimension is still too high, the detection accuracy is reduced too much, and so on.
In order to reduce the feature dimension of DCTR without affecting the detection of the stego images, this paper attempts to propose a DCTR feature selection method based on different criteria with gain-loss (abbreviated as S-FUND method). First, we try to give a measurement algorithm based on distortion and information gain ratio according to which we measure the difference between the covers and stegos for each steganalysis feature component. Second, by setting the threshold, we delete the feature components with a small contribution measured by at least one of the two criteria. Finally, the preserved feature components are used as the final selected steganalysis feature for training and detection. This method is expected to reduce the spatial complexity of detecting the stego images by reducing the feature dimension while maintaining or even improving the detection accuracy, and to reduce the time complexity of detecting the stego images by avoiding the dependence on classification results.
The rest of this paper is organized as follows:
Section 2 describes the related work.
Section 3 outlines a measurement algorithm for the difference of steganalysis feature components.
Section 4 proposes a multi-criteria-based feature selection method.
Section 5 analyzes the effect of DCTR features based on S-FUND method proposed in this paper for detecting of the stego images through a series of experimental comparisons.
Section 6 further discusses the experiment in this paper. Finally,
Section 7 summarizes the whole paper.
2. Related Work
Distortion is a measure of the difference between the original signal and the altered signal [
38], and it is defined as the square root of the ratio of the total harmonic energy to the fundamental energy by the following formula.
where
and
P represent the fundamental energy and the total signal energy, respectively, and
represents the energy of the
nth harmonic. The larger the
K value, the greater the ratio of total harmonic energy to fundamental energy, i.e., the greater the difference between harmonic energy and fundamental energy, and thus the better to distinguish harmonic energy from fundamental energy.
The information gain ratio improves the measure of the difference between features [
39], and solves the problem of information gain bias to take more features, which is defined as the ratio of the information gain value to the entropy
of the value of dataset
X for feature
Y, with the following formula.
where
represents the information gain value between dataset
X and feature
Y,
and
represent the entropy values of training dataset
X and feature
Y, respectively,
represents the uncertainty of
X given known
Y.
,
n is the number of
Y values,
represents the number of samples in the
X. The smaller the value of
, the smaller the degree of uncertainty in
Y decreasing with
X, i.e., the smaller the difference between them.
In previous research, researchers have often used one criterion first to select some of the features that are useful for detecting the stego images (written as “First Selection”), and then using another criterion for a second selection based on the “First Selection” [
33,
34,
35,
36,
37]. Although this does reduce the feature dimension, the two criteria are used sequentially, i.e., one criterion is measured ahead of the other, so that there are inconsistencies in the measurement using the two criteria.
For example, Yu et al. [
37] proposed the SRGS method to reduce the feature dimension. First, the effective SNR criterion is improved, and valuable feature components are selected. On this basis, the Relief algorithm is improved, the redundant feature components in the remaining feature components are deleted. Afterwards, the valuable and non-redundant feature components are finally selected. However, in the SRGS method, when using the Relief algorithm, some components retained by the SNR criterion are unable to be measured by other criteria. There is the same dilemma in [
34,
35]. This requires the reliability of the selected features to be further improved.
Therefore, it is necessary to find a more rigorous method for selecting valuable feature components to further reduce the feature dimension and improve the detection accuracy.
4. S-FUND Method
Based on the above, this paper proposes an S-FUND method, which treats the two criteria for measuring the difference between the cover and stego features as the same. First, two criteria are used to measure the difference between the cover and stego features. Afterwards, the feature component with the larger difference in order is deleted. Finally, the preserved feature components are used as the final feature.
4.1. Specific Algorithm
We present the specific algorithm of S-FUND method according to the main steps described above, as shown in Algorithm 1.
Algorithm 1 Specific algorithm of S-FUND method |
- Input:
The original feature ; - Output:
Final selection of steganalysis feature based on S-FUND method , where m is the final feature dimension; - 1:
for to N do - 2:
for to n do - 3:
Using Equation ( 5) to measure the sum of the ith feature component differences between the cover and stego images (written as ); - 4:
Using step 3 and Equation ( 5) ⇒ Computing the value of the ith feature component between the cover and stego images; - 5:
end for - 6:
end for - 7:
Feature components arranged in descending order according to value in step 4 ; - 8:
for to N do - 9:
for to n do - 10:
Using Equation ( 7) to calculate the value of the feature component in the cover images with respect to it in the stego images; - 11:
According to the value from step 10, using Equation ( 6) ⇒ Calculating the value of the feature component between the cover and stego images; - 12:
Using Equation ( 9) to calculate the value of the feature component in the cover images with respect to the value of the feature component in the stego images; - 13:
According to the value in step 11, the value in step 12, using Equation ( 8) ⇒ Calculating the value of the feature component between the cover and stego features; - 14:
end for - 15:
end for - 16:
According to the value in step 13, steganalysis feature components in descending order ; - 17:
Setting threshold T; - 18:
for to N do - 19:
Sorting results according to steps 7 and 16 ⇒ Calculating the difference between the kth feature component ranked according to the two criteria (written as ); - 20:
end for - 21:
Deleting the feature components which the absolute value of the ranking difference is greater than the threshold T;
- 22:
Preserving the m-dimensional eligible component for training and detection as the final selected feature.
|
Thus, in this paper, we measure the difference between the cover and stego features using distortion factor and information gain ratio. This method reduces the feature dimension significantly, thus reducing the spatio-temporal complexity of detecting the stego images.
Next,
Figure 1 is given, which is based on Algorithm 1, and visually depicts S-FUND method process to select the steganalysis features.
In the next subsection, we analyze the time complexity of each of the main steps in S-FUND method and compare it with the time complexity of the classification results relying on Fisher’s linear discriminant integrated classifier to give the readers a better understanding of the performance of this method.
4.2. Performance Analysis
The time complexity of the main steps of the proposed S-FUND method is analyzed separately, and the time complexity of other existing classic methods are compared, as shown in
Table 1.
There is no nested relationship between the steps in
Table 1, so the time complexity of the S-FUND method proposed in this paper is equal to the maximum time complexity of all the steps. When
, i.e.,
, the time complexity of the S-FUND method is
; when
, the time complexity of the S-FUND method is
. However, existing feature selection methods rely on the classification results of the Fisher linear discriminant integrated classifier [
40], which has a time complexity of:
where
L represents the number of individual learners,
represents number of training sets per type,
represents subspace dimension, so the time complexity of this class of selection methods is
must be greater than or equal to
, i.e.,
. Thus, the time complexity of the selection method that relies on Fisher Linear Discriminant integrated classifier results is much greater than
or
. Because the DCTR feature dimension is 8000,
and
. The time complexity of S-FUND method is less than that of PCA-D, Steganalysis-
, Fisher-G and SRGS methods, and it is similar to that of the CGSM method.
Therefore, the S-FUND method greatly reduces the running time complexity and improves the efficiency of detecting the stego images.
4.3. Threshold Analysis
In order to have a better selection of S-FUND method, we need to explain the threshold T in Algorithm 1.
Based on previous work and experimental experience, when the DCTR feature is reduced to about 40% of the original by different methods, they all show worse detection effect than the original, and the detection effect will decrease as the feature dimension decreases. In this method, when the feature dimension is reduced to about 40% of the original, the threshold is 0.15.
Therefore, we initially set , if is greater than T, that is, , it means that the contribution degree of the feature components measured by the two criteria is quite different. Thus, the two feature components are removed to reduce the feature dimension. We are able to obtain the value of T, i.e., by setting the steps of 0.01. Then, by comparing the detection accuracy under these thresholds T, S-FUND method with the highest detection accuracy is selected as the final detection effect under this payload. In this way, more valuable feature components can be selected, and the feature dimension will be greatly reduced without affecting the detection accuracy, so as to achieve the purpose of feature selection for steganalysis.
5. Experimental Results and Analysis
In order to detect the performance of the S-FUND method proposed in this paper, a series of selection and comparison experiments were conducted using 8000-D DCTR feature [
27]. All experiments were run in MATLAB R2018a with an Intel i7-8550U CPU and an 8G RAM computer to ensure that the different methods could be fairly compared. Experimental result figures were processed and generated in OriginPro 8.5.
5.1. Experimental Setup
The computer software, hardware, image library and steganalysis features used in all the experiments in this paper are the same to ensure that the different methods can be compared fairly and more reliably.
We performed a series of operations on the BOSSbase 1.01 image library of the website (BOSSbase 1.01 image library from the URL:
http://dde.binghamton.edu/download/, accessed on 3 March 2020) to prepare for the next experiments, with the following steps.
(1) Converting 10,000 images in PGM format from the BOSSbase 1.01 image library into JPEG images with a compression quality factor of 95 (QF = 95).
(2) Generating 10,000 JPEG cover images using SI-UNIWARD [
6] steganography to generate 10,000 × 5 = 50,000 images with payload of 0.1, 0.2, 0.3, 0.4 and 0.5 (bpAC).
(3) The DCTR [
27] extraction algorithm was used to extract 8000-D steganalysis features from the cover and stego images, and 10,000 × (1 + 5) = 60,000 steganalysis features were obtained. The reason for this is that the cover images become stego images after embedding information, training and detection are paired, that is, the images are symmetric.
Steganalysis effect depends not only on the types of steganalysis features, but also on the different steganography schemes and payloads. Early steganography schemes can be accurately detected by high-dimensional steganalysis features (detection accuracy can reach more than 80%), such as LSB, nsF5 [
1], and so on. In recent years, adaptive steganography has achieved low embedding jitter, such as SI-UNIWARD [
6]. Solving the ones difficult to detect is the focus of this paper. Moreover, the detection accuracy of the same steganalysis features will be higher with the increase of payloads.
Meanwhile, the steganographic images with high payloads can be visually distinguished to a certain extent, which loses the significance of steganalysis, thus, the payloads of this paper are 0.1, 0.2, 0.3, 0.4 and 0.5. The specific experimental object settings are shown in
Table 2.
Kodovský et al. [
40] proposed the FLD ensemble classifier based on random forest algorithm and Monte Carlo thought, and it is widely used to train and detect for steganalysis feature selection [
33,
34,
35,
36,
37]. Therefore, the FLD integrated classifier is used in this paper, which can achieve the purpose of fair comparison between different methods. Firstly, one-half of the cover image features and their corresponding stego image features with different payloads are randomly selected from each feature set as the training set. Afterwards, the remaining cover image features and their corresponding stego image features with different payloads are used as the test set. The error rates in this integrated classifier are calculated as
where
and
represent the false alarm rate and missed detection rate, respectively,
represents the number of test sets, because the test set contains a cover image set and a stego image set, i.e.,
. The error rate represents the ratio of the number of classification errors to the total number of tested feature components. The lower the detection error rate, the better the selected feature components are at detecting the stego images. In order to show more intuitively the results of the comparison experiment, we use the following equation—
to convert the detection error rate obtained by the classifier into detection accuracy.
represents average detection accuracy—the larger
value is, the better the selected features are for detecting the stego images.
The experiment consisted of four parts:
- 1.
Comparison experiment based on the S-FUND method before and after the selection of DCTR steganalysis feature [
27] (
Section 5.2);
- 2.
Comparison experiment with Random-D method (
Section 5.3);
- 3.
- 4.
5.2. Selection Experiment
Holub et al. [
27] proposed an 8000-D DCTR image steganalysis feature. The feature is a first-order statistic of the quantized noise residuals obtained from the decompressed JPEG image using 64 discrete Cosine transform kernels. The DCTR feature has lower dimension and computational complexity, and better detection performance.
In order to obtain the detection accuracy of the S-FUND method, we set T = 0.15, 0.14, 0.13, …, 0.02 and 0.01 by setting the steps of 0.01, if is greater than T, the two feature components are removed to reduce the feature dimension.
In the S-FUND method, we delete the steganalysis feature components that differ more than the threshold values measured by two criteria. Firstly, in order to effectively reduce the feature dimension, we consider the difference between the two criteria greater than 15% of the original feature dimensions to be large, and for 8000-D DCTR feature, we delete the component of steganalysis features with a difference greater than 8000 × 15% = 1200 dimensions. The experimental results for the selected feature dimensions and detection accuracy are shown in
Table 3.
In
Table 3, ‘Dim’ represents the feature dimension, and represents the detection accuracy. From
Table 3, it can be seen that the S-FUND method can significantly reduce the feature dimension while maintaining or even improving the detection accuracy at different payloads. For example, when payload = 0.1, the selected features based on S-FUND method can achieve a detection accuracy of 0.5270, which is 0.31% higher than the original, and the selected feature is 3462 dimensions lower than the original feature dimension. Moreover, when
, while maintaining the detection accuracy of the stego images, the feature dimension selected based on S-Fund method is only 30.44% of the original. When payload = 0.2, 0.3, the features selected based on S-FUND method can all reduce the DCTR feature dimension to different degrees, and the detection accuracy is improved by 0.49% and 0.16%, respectively. Moreover, the features selected based on S-FUND method are only 29.79% and 51.79% of the original while maintaining the detection accuracy of the stego images, which reduces the spatio-temporal cost of classifier training.
In order to compare the selection of DCTR steganalysis feature by S-FUND method more visually, the feature dimension and detection accuracy before and after selection are shown in
Figure 2 below.
In
Figure 2, the horizontal axis represents threshold value and the vertical axis represents the corresponding feature dimension and detection accuracy, and the five lines from top to bottom represent the effects of the DCTR features selected at five different payloads, and the points of optimal performance at each payload are processed and labeled with values. It can be clearly seen from the figure that the S-FUND method can maintain or even improve the detection accuracy of DCTR feature while significantly reducing the feature dimension, which proves the effectiveness of the S-FUND method.
5.3. Comparison Experiment with Random-D Method
In order to make a fair comparison between the different methods, the experimental setup of this comparison experiment is the same as in
Section 5.1. For different payloads, the following is a comparison of the detection accuracy of the selected features of S-FUND method and the selected features of Random-D method for the stego images. We let the feature dimensions selected by the Random-D method be equal to the corresponding feature dimensions selected by S-FUND method, and compare their detection accuracy; the comparison results are shown in
Table 4.
It can be seen from
Table 4 that the S-FUND method has better detection accuracy than Random-D method when DCTR feature is reduced to the same dimension at different payloads. For example, when payload = 0.3, the detection accuracy of the features selected by the S-FUND method is 0.52%, 0.61%, 0.83%, 0.97%, 0.98% and 0.98% higher than that of the Random-D method by reducing the feature dimension to 5653, 4745, 3801, 2750, 1826 and 848 dimensions, respectively.
In order to compare the effects of S-FUND method and Random-D method on the selection of DCTR feature more visually, we created
Figure 3 based on
Table 4.
In
Figure 3, the horizontal axis represents the number of feature dimensions, and the vertical axis represents the corresponding detection accuracy. Five line graphs from top to bottom represent the effects of DCTR features selected by S-FUND and Random-D selection at five different payloads. From
Figure 3, we can see that the S-FUND method has better detection accuracy than the Random-D method at different payloads.
Additionally, we found that as the number of selected features decreases, the difference in detection accuracy between the two methods of the stego images increases. For example, when payload = 0.1, the difference between the two methods increases from 0.32% to 1.3%. The reason for this phenomenon may be the following: as the number of selected features decreases, the useful features will more likely be deleted by Random-D, while the useless features will be deleted by S-FUND. Therefore, as the number of selected decreases, the difference between the detection accuracy of the stego images based on the features selected by S-FUND method and Random-D selection method will increase.
5.4. Comparison Experiment with CGSM Method
Wang et al. [
35] proposed a method for steganalysis feature selection based on the difference function and Pearson’s correlation coefficient. The method first presents a difference function-based feature measurement algorithm to measure the difference between the cover and stego features. Afterwards, the Pearson correlation coefficient is improved to measure the correlation between the feature components and the image classification result. Finally, the feature component with large difference function is selected and the feature component with small Pearson correlation coefficient is deleted. This method is able to select the effective feature components to improve the detection performance of the stego images, however, there is a sequence between the two algorithms, which makes it difficult to avoid selecting useless features or deleting valuable features.
For the DCTR steganalysis feature with different payloads, the following is a comparison between the S-FUND method and the CGSM method when the same feature dimensions are selected, and the comparison results are shown in
Table 5.
It can be seen from
Table 5 that S-FUND method has better detection accuracy than CGSM method when the selected DCTR feature is reduced to the same dimensions by both the S-FUND method and the CGSM method at different payloads. For example, when payload = 0.1, the detection accuracy of selected features of the stego images by S-FUND method is 0.12%, 0.43%, 0.56%, 0.89%, 1.16% and 1.6% higher than that of CGSM method when the feature dimension is reduced to 5791, 4826, 3936, 2872, 2015 and 849.
In order to compare the selection effects of S-FUND and CGSM methods on the DCTR steganalysis features more visually, we made
Figure 4 based on
Table 5.
In
Figure 4, the horizontal axis represents the feature dimension and the vertical axis represents the corresponding detection accuracy. Five line graphs from top to bottom represent the effects of S-FUND and CGSM methods for selecting DCTR feature at five different payloads. From
Figure 4, it can be seen that the S-FUND method has better detection accuracy than the CGSM method when both methods are reduced to the same dimensions under different payloads. Simultaneously, as the number of selected features decreases at different payloads, the difference in detection accuracy between the two methods increases—for example, when payload = 0.1, the difference in detection accuracy between the two methods increases from 0.12% to 1.6%.
Furthermore, we found that the difference in detection accuracy of the stego images between the features selected by the two selection methods is greater at high payloads. For example, when payload = 0.1, 0.2, 0.3, 0.4 and 0.5, the differences are 1.6%, 1.6%, 1.85%, 2.25% and 1.87%, respectively. The reason for this phenomenon may be that more information is embedded in the high payload, and the difference between the cover and stego features is larger, which makes it easier to detect the stego images with the features selected by S-FUND method. Therefore, at high payloads, the difference in detection accuracy of the stego images with the features selected by S-FUND method is larger than that of CGSM method.
5.5. Comparison Experiment with the PCA-D Method
Qin et al. [
33] proposed the PCA-D method which has no parameter restrictions, is simple to use, and the results are easier to understand. However, the method is not efficient and is less effective for data with nonlinear structure. The main process is: Firstly, each feature component is subtracted from its respective mean. Secondly, the covariance matrix and the eigenvalues and eigenvectors of the covariance matrix are calculated. Thirdly, the feature values are sorted in descending order. Finally, the number of principal components (the number of selected feature components) is determined, and the feature component is selected as the final selected feature vector, and the comparison experimental results between the S-FUND and the PCA-D methods are shown in
Table 6.
As can be seen from
Table 6, the DCTR features selected by S-FUND methods have better detection accuracy than those selected by PCA-D method for the cover images when the DCTR feature is reduced to the same dimensions at different payloads by the two methods. For example, when payload = 0.2, the detection accuracy of S-FUND method is 0.02%, 1.83%, 0.95% and 1.2% higher than that of the PCA-D method when the feature is reduced to 5739, 4790, 3870 and 2790 dimensions, respectively. When payload = 0.4, using the two methods, the detection accuracy of the features selected by S-FUND method is 2.39%, 2.85%, 1.79% and 0.87% higher than that of the PCA-D method by reducing the feature dimensions to 5531, 4608, 3690 and 2664 dimensions, respectively.
In order to compare more intuitively the effects of the selected features of S-FUND and PCA-D methods for DCTR feature, we created
Figure 5 based on
Table 6.
In
Figure 5, the horizontal axis represents the feature dimension, the vertical axis represents the corresponding detection accuracy, and the five line graphs from top to bottom represent the effects of S-FUND and PCA-D methods for selecting DCTR feature at five different payloads. From
Figure 5, it can be seen that the S-FUND method has better detection accuracy than the PCA-D method when the two methods are reduced to the same dimension at different payloads. For example, when payload = 0.1, the feature dimensions selected by the two methods are 5791 and 2435, and the differences in detection accuracy are 1.54% and 0.95%. When payload = 0.3, the feature dimensions selected by the two methods are 5653 and 2280 and the differences in detection accuracy are 2.00% and 0.51%. When payload = 0.5, the feature dimensions selected by the two methods are 5192 and 2621, and the differences in detection accuracy are 4.25% and 1.23%.
6. Discussion
The S-FUND method has also been used in other steganalysis feature selection experiments, such as CC-JRM. Therefore, the method in this paper is a general feature selection method for steganalysis. It is worth mentioning that the contribution of this paper is not only the introduction of distortion and information gain ratio to measure the difference of feature components between the cover and stego images, but also the idea that the two criteria are considered at the same time.
Last but not least, since this work considers the two criteria at the same time, they need to measure the same kind of criteria, that is, they either measure the contribution degree, or both measure the degree of similarity, or both measure redundancy, and so on. The different kinds of criteria in this task force have certain limitations. Regarding the CGSM method and the SRGS method, one involves measuring the similarity between the feature components, and the other measuring the uselessness of the feature component. Such criteria cannot be applied by this paper. In the next work, we are expected to propose a novel method for different kinds of standards. When the kinds of criterion are different, the combination of the these can be carried out—this method considers to combine the similarity between feature components and the value of the component. Among the two feature components with similarity, the component with a small contribution is deleted.
7. Conclusions
In order to effectively reduce the number of DCTR steganalysis feature while maintaining or even improving the detection accuracy of the stego images, this paper proposes a feature selection approach based on distortion factor and information gain ratio. First, the distortion factor and information gain ratio are improved to measure the difference of each steganalysis feature component between the cover and stego images. Second, two measurement values of each feature component are arranged in descending order. Based on the above, the feature components with large different ranking according to the two measurements are deleted. Finally, the preserved feature components are used as the final feature vector. This approach can effectively reduce the DCTR feature dimension while maintaining or even improving the detection accuracy of the stego images, thus reducing the spatial complexity of detecting the stego images. Moreover, by comparing the time complexity of S-FUND method and the selection method that relies on the FLD classifier results, it is demonstrated that the approach in this paper can greatly improve the operational efficiency, thus reducing the time complexity of the classifier to detect the stego images and reducing the cost of detection.
We have conducted a large number of comparison experiments to indicate that DCTR feature selected by S-FUND method substantially reduce the feature dimension while maintaining or even improving the detection accuracy of the stego images. And then, we compared S-FUND method with Random-D, CGSM, and PCA-D methods to indicate that the features selected by S-FUND method are more detection accurate of the stego images. For example, in the comparison experiments with PCA-D method, the features selected by the S-FUND method up to 4.25% more higher than PCA-D method for the stego images.