Abstract
In order to improve the feature selection stability based on evolutionary algorithms, an evolutionary algorithms’ feature selection stability improvement system is proposed. Three Filter methods’ results are aggregated to provide the stability information, and feature selection stability and classification accuracy are adopted as two optimization objectives. Weighted sum, weighted product and biobjective optimization methods together are applied as the system’s optimization models. Ant colony optimization, particle swarm optimization and genetic algorithm are used as testing algorithms, and experiments are taken on two benchmark datasets. The results show that the proposed system can improve the stability of evolutionary algorithms’ feature selection efficiently and their classification performance simultaneously.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Emani, C.K., Cullot, N., Nicolle, C.: Understandable big data: a survey. Comput. Sci. Rev. 17, 70–81 (2015)
OSullivan, B., Wooldridge, M.: Feature Selection for High Dimensional Data. Springer, Heidelberg (2015)
Guo, H.X., Li, Y.J., Shang, J., Gu, M.Y., Huang, Y.Y., Gong, B.: Learning from class imbalanced data: review of methods and applications. Expert. Syst. Appl. 73, 220–239 (2017)
Fan, M., Chou, C.A.: Exploring stability based voxel selection methods in mvpa using cognitive neuroimaging data: a comprehensive study. Brain Inform. 3, 193–203 (2016)
Kalousis, A., Prados, J., Hilario, M.: Stability of feature selection algorithms: a study on high-dimensional spaces. Knowl. Inf. Syst. 12, 95–116 (2007)
Garcia-Torres, M., Gomez-Vela, F., Melian-Batista, B., Moreno-Vega, J.M.: High-dimensional feature selection via feature grouping: a variable neighborhood search approach. Inf. Sci. 326, 102–118 (2016)
Li, Y., Si, J., Zhou, G.J., Huang, S.S., Chen, S.C.: FREL: a stable feature selection algorithm. IEEE Trans. Neural Netw. Learn. Syst. 26, 1388–1402 (2015)
Somol, P., Novovicovaa, J.: Evaluating stability and comparing output of feature selectors that optimize feature subset cardinality. IEEE Trans. Pattern Anal. 32, 1921–1939 (2010)
Tohka, J., Moradi, E., Huttunen, H.: Comparison of feature selection techniques in machine learning for anatomical brain MRI in dementia. Neuroinformatics 14, 1–18 (2016)
Zhou, Q.F., Ding, J.C., Ning, Y.P., Luo, L.K., Li, T.: Stable feature selection with ensembles of multi-reliefF. In: 10th International Conference on Natural, pp. 742–747. IEEE Press, New York (2014)
Fahad, A., Tari, Z., Khalil, I., Almalawi, A.Y., Zomaya, A.: An optimal and stable feature selection approach for traffic classification based on multi-criterion fusion. Future Gener. Comput. Syst. 36, 156–169 (2014)
Kim, H.J., Choi, B.S., Huh, M.Y.: Booster in high dimensional data classification. IEEE Trans. Knowl. Data Eng. 28, 29–40 (2016)
Pes, B., Dessi, N., Angioni, M.: Exploiting the ensemble paradigm for stable feature selection: a case study on high-dimensional genomic data. Inf. Fusion 35, 132–147 (2017)
Wang, H., Khoshgoftaar, T.M., Seliya, N.: On the stability of feature selection methods in software quality prediction: an empirical investigation. Int. J. Softw. Eng. Know. 25, 1467–1490 (2015)
Yu, L., Ding, C., Loscalzo, S.: Stable feature selection via dense feature groups. In: 14th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pp. 803–811. ACM, New York (2008)
Kamker, I., Gupta, S.K., Phung, D., Venkatesh, S.: Stabilizing \(l_1\)-norm prediction models by supervised feature grouping. J. Biomed. Inform. 59, 149–168 (2016)
Shu, L., Ma, T.Y., Latecki, L.J.: Stable feature selection with minimal independent dominating sets. In: ACM International Conference on Bioinformatics, pp. 450–457. ACM, New York (2013)
Beinrucker, A., Dogan, U., Blanchard, G.: Extensions of stability selection using subsamples of observations and covariates. Stat. Comput. 5, 1–19 (2016)
Erguzel, T.T., Ozekes, S., Gultekin, S., Tarhan, N.: Ant colony optimization based feature selection method for QEEG data classification. Psychiatr. Invest. 11, 243–250 (2014)
Singh, S., Selvakumar, S.: A hybrid feature subset selection by combining filters and genetic algorithm. In: International Conference on Computing. Communication and Automation, pp. 283–289. IEEE Press, New York (2015)
Dudek, G.: Artificial immune system with local feature selection for short term load forecasting. IEEE Trans. Evol. Comput. 21, 116–130 (2017)
Xue, B., Zhang, M.J., Brownw, W.N., Yao, X.: A survey on evolutionary computation approaches to feature selection. IEEE Trans. Evol. Comput. 20, 606–626 (2016)
Zhang, Y., Gong, D.W., Cheng, J.: Multiobjective particle swarm optimization approach for cost based feature selection in classification. IEEE ACM Trans. Comput. Bioinform. 14, 64–75 (2017)
Aldehim, G., Wang, W.J.: Weighted heuristic ensemble of filters. In: SAI Intelligent Systems Conference, pp. 609–615. IEEE Press, New York (2015)
Nogueira, S., Brown, G.: Measuring the stability of feature selection with applications to ensemble methods. In: Schwenker, F., Roli, F., Kittler, J. (eds.) MCS 2015. LNCS, vol. 9132, pp. 135–146. Springer, Cham (2015). https://doi.org/10.1007/978-3-319-20248-8_12
Acknowledgments
This work was supported by the Natural Science Foundation of China under Grant 61371196.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2017 Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Liu, Y., Diao, X., Cao, J., Zhang, L. (2017). Evolutionary Algorithms’ Feature Selection Stability Improvement System. In: He, C., Mo, H., Pan, L., Zhao, Y. (eds) Bio-inspired Computing: Theories and Applications. BIC-TA 2017. Communications in Computer and Information Science, vol 791. Springer, Singapore. https://doi.org/10.1007/978-981-10-7179-9_6
Download citation
DOI: https://doi.org/10.1007/978-981-10-7179-9_6
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-10-7178-2
Online ISBN: 978-981-10-7179-9
eBook Packages: Computer ScienceComputer Science (R0)