Nothing Special   »   [go: up one dir, main page]

Skip to main content

Advertisement

Log in

Double-kernelized weighted broad learning system for imbalanced data

  • Original Article
  • Published:
Neural Computing and Applications Aims and scope Submit manuscript

Abstract

Broad learning system (BLS) is an emerging neural network with fast learning capability, which has achieved good performance in various applications. Conventional BLS does not effectively consider the problems of class imbalance. Moreover, parameter tuning in BLS requires much effort. To address the challenges mentioned above, we propose a double-kernelized weighted broad learning system (DKWBLS) to cope with imbalanced data classification. The double-kernel mapping strategy is designed to replace the random mapping mechanism in BLS, resulting in more robust features while avoiding the step of adjusting the number of nodes. Furthermore, DKWBLS considers the imbalance problem and achieves more explicit decision boundaries. Numerous experimental results show the superiority of DKWBLS in tackling imbalance problems over other imbalance learning approaches.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4

Similar content being viewed by others

Explore related subjects

Discover the latest articles, news and stories from top researchers in related subjects.

References

  1. Gong W, Zhang W, Bilal M, Chen Y, Xu X, Wang W (2022) Efficient Web APIs Recommendation with privacy-preservation for mobile app development in industry 4.0. IEEE Trans Ind Info. https://doi.org/10.1109/TII.2021.3133614

    Article  Google Scholar 

  2. Deebak BD, Memon FH, Khowaja SA, Dev K, Wang W, Qureshi NMF, Su C (2022) Lightweight Blockchain Based Remote Mutual Authentication for AI-Empowered IoT sustainable computing systems. IEEE Intern Things J. https://doi.org/10.1109/JIOT.2022.3152546

    Article  Google Scholar 

  3. Deebak BD, Memon FH, Dev K, Khowaja SA, Wang W, Qureshi NMF (2022) TAB-SAPP: a trust-aware blockchain-based seamless authentication for massive IoT-enabled industrial applications. IEEE Trans Indust Info. https://doi.org/10.1109/TII.2022.3159164

    Article  Google Scholar 

  4. Haixiang G, Yijing L, Shang J, Mingyun G, Yuanyue H, Bing G (2017) Learning from class-imbalanced data: review of methods and applications. Expert Sys Appl 73:220–239

    Article  Google Scholar 

  5. Sarmanova A, Albayrak S (2013) Alleviating class imbalance problem in data mining. In 2013 21st Signal Processing and Communications Applications Conference (SIU) (pp. 1-4), April. IEEE

  6. Alsolai H, Roper M (2020) A systematic literature review of machine learning techniques for software maintainability prediction. Info Softw Tech 119:106214

    Article  Google Scholar 

  7. Malhotra R (2015) A systematic review of machine learning techniques for software fault prediction. Appl Soft Comput 27:504–518

    Article  Google Scholar 

  8. Cui L, Yang S, Chen F, Ming Z, Lu N, Qin J (2018) A survey on application of machine learning for internet of things. Int J Mach Learn Cybern 9(8):1399–1417

    Article  Google Scholar 

  9. Krawczyk B, Galar M, Jeleń Ł, Herrera F (2016) Evolutionary undersampling boosting for imbalanced classification of breast cancer malignancy. Appl Soft Comput 38:714–726

    Article  Google Scholar 

  10. Wei W, Li J, Cao L, Ou Y, Chen J (2013) Effective detection of sophisticated online banking fraud on extremely imbalanced data. World Wide Web 16(4):449–475

    Article  Google Scholar 

  11. Zakaryazad A, Duman E (2016) A profit-driven Artificial Neural Network (ANN) with applications to fraud detection and direct marketing. Neurocomputing 175:121–131

    Article  Google Scholar 

  12. Wang K, Liu L, Yuan C, Wang Z (2021) Software defect prediction model based on LASSO\(\ddot{\,}\)CSVM. Neur Comput Appl 33(14):8249–8259

    Article  Google Scholar 

  13. Lee T, Nam J, Han D, Kim S, In HP (2016) Developer micro interaction metrics for software defect prediction. IEEE Trans Softw Eng 42(11):1015–1035

    Article  Google Scholar 

  14. Krawczyk B (2016) Learning from imbalanced data: open challenges and future directions. Progr Artif Intell 5(4):221–232

    Article  Google Scholar 

  15. Hsiao YH, Su CT, Fu PC (2020) Integrating MTS with bagging strategy for class imbalance problems. Int J Mach Learn Cybern 11(6):1217–1230

    Article  Google Scholar 

  16. Zhe W et al (2020) Multi-matrices entropy discriminant ensemble learning for imbalanced problem. Neural Comput Appl 32(12):8245–8264

    Article  Google Scholar 

  17. Liu XY, Wu J, Zhou ZH (2008) Exploratory undersampling for class-imbalance learning. IEEE Trans Sys, Man and Cybern, Part B (Cybernetics) 39(2):539–550

    Google Scholar 

  18. Elkan C (2001) The foundations of cost-sensitive learning. In International joint conference on artificial intelligence (Vol. 17, No. 1, pp. 973-978). Lawrence Erlbaum Associates Ltd

  19. Zhou ZH, Liu XY (2005) Training cost-sensitive neural networks with methods addressing the class imbalance problem. IEEE Trans Knowl Data Eng 18(1):63–77

    Article  Google Scholar 

  20. Chen CP, Liu Z (2017) Broad learning system: an effective and efficient incremental learning system without the need for deep architecture. IEEE Trans Neural Netw Learn Systems 29(1):10–24

    Article  MathSciNet  Google Scholar 

  21. Zhang L et al (2020) Analysis and variants of broad learning system. IEEE Trans Sys, Man Cybern Sys. https://doi.org/10.1109/TSMC.2020.2995205

    Article  Google Scholar 

  22. LeCun Y, Huang FJ, Bottou L (2004) Learning methods for generic object recognition with invariance to pose and lighting. In Proceedings of the 2004 IEEE Computer society conference on computer vision and pattern recognition, 2004. CVPR 2004. (Vol. 2, pp. II-104). IEEE

  23. Yi H, Shiyu S, Xiusheng D, Zhigang C (2016) A study on deep neural networks framework. In 2016 IEEE Advanced information management, communicates, electronic and automation control conference (IMCEC) (pp. 1519-1522). IEEE

  24. Chawla NV, Bowyer KW, Hall LO, Kegelmeyer WP (2002) SMOTE: synthetic minority over-sampling technique. J Artif Intell Res 16:321–357

    Article  MATH  Google Scholar 

  25. Han X, Cui R, Lan Y, Kang Y, Deng J, Jia N (2019) A Gaussian mixture model based combined resampling algorithm for classification of imbalanced credit data sets. Int J Mach Learn Cybern 10(12):3687–3699

    Article  Google Scholar 

  26. Han H, Wang WY, Mao BH (2005) Borderline-SMOTE: a new over-sampling method in imbalanced data sets learning. In: International conference on intelligent computing (pp. 878-887). Springer, Berlin and Heidelberg

  27. Bunkhumpornpat C, Sinapiromsaran K, Lursinsap C (2009). Safe-level-smote: safe-level-synthetic minority over-sampling technique for handling the class imbalanced problem. In Pacific-Asia conference on knowledge discovery and data mining (pp. 475-482), April. Springer, Berlin and Heidelberg

  28. Maciejewski T, Stefanowski J (2011, April) Local neighbourhood extension of SMOTE for mining imbalanced data. In 2011 IEEE symposium on computational intelligence and data mining (CIDM) (pp. 104-111). IEEE

  29. Sauptik D, Vladimir C (2014) Development and evaluation of cost-sensitive universum-SVM. IEEE Trans Cybern 45(4):806–818

    Google Scholar 

  30. Hazarika BB, Deepak Gupta (2021) Density-weighted support vector machines for binary class imbalance learning. Neural Comput Appl 33(9):4243–4261

    Article  Google Scholar 

  31. Xu R, Wen Z, Gui L, Lu Q, Li B, Wang X (2020) Ensemble with estimation: seeking for optimization in class noisy data. Int J Mach Learn Cybern 11(2):231–248

    Article  Google Scholar 

  32. Yong Z et al (2017) Ensemble weighted extreme learning machine for imbalanced data classification based on differential evolution. Neural Comput Appl 28(1):259–267

    Google Scholar 

  33. Krawczyk B, Woźniak M, Schaefer G (2014) Cost-sensitive decision tree ensembles for effective imbalanced classification. Appl Soft Comput 14:554–562

    Article  Google Scholar 

  34. Xiao W, Zhang J, Li Y, Zhang S, Yang W (2017) Class-specific cost regulation extreme learning machine for imbalanced classification. Neurocomputing 261:70–82

    Article  Google Scholar 

  35. Raghuwanshi BS, Shukla S (2018) Class-specific extreme learning machine for handling binary class imbalance problem. Neural Netw 105:206–217

    Article  MATH  Google Scholar 

  36. Zong W, Huang GB, Chen Y (2013) Weighted extreme learning machine for imbalance learning. Neurocomputing 101:229–242

    Article  Google Scholar 

  37. Yang K, Yu Z, Chen CP, Cao W, You JJ, San Wong H (2021) Incremental weighted ensemble broad learning system for imbalanced data. IEEE Trans Knowl Data Eng. https://doi.org/10.1109/TKDE.2021.3061428

    Article  Google Scholar 

  38. Raghuwanshi BS, Shukla S (2021) Classifying imbalanced data using SMOTE based class-specific kernelized ELM. Int J Mach Learn Cybern 12(5):1255–1280

    Article  Google Scholar 

  39. Xiaokang W, Huiwen W, Yihui W (2020) A density weighted fuzzy outlier clustering approach for class imbalanced learning. Neural Comput Appl 32(16):13035–13049

    Article  Google Scholar 

  40. Mao W, Wang J, Xue Z (2017) An ELM-based model with sparse-weighting strategy for sequential data imbalance problem. Int J Mach Learn Cybern 8(4):1333–1345

    Article  Google Scholar 

  41. Kemal P (2018) Similarity-based attribute weighting methods via clustering algorithms in the classification of imbalanced medical datasets. Neural Comput Appl 30(3):987–1013

    Article  Google Scholar 

  42. Georgios D, Fernando B, Felix L (2018) Improving imbalanced learning through a heuristic oversampling method based on k-means and SMOTE. Info Sci 465:1–20

    Article  Google Scholar 

  43. Pao YH, Park GH, Sobajic DJ (1994) Learning and generalization characteristics of the random vector functional-link net. Neurocomputing 6(2):163–180

    Article  Google Scholar 

  44. Igelnik B, Pao YH (1995) Stochastic choice of basis functions in adaptive function approximation and the functional-link net. IEEE Trans Neural Netw 6(6):1320–1329

    Article  Google Scholar 

  45. Olshausen BA, Field DJ (1997) Sparse coding with an overcomplete basis set: a strategy employed by V1? Vision Research 37(23):3311–3325

    Article  Google Scholar 

  46. Zhang L, Suganthan PN (2017) Benchmarking ensemble classifiers with novel co-trained kernel ridge regression and random vector functional link ensembles [research frontier]. IEEE Comput Intell Magaz 12(4):61–72

    Article  Google Scholar 

  47. Yu Z, Lan K, Liu Z, Han G (2021) Progressive ensemble kernel-based broad learning system for noisy data classification. IEEE Trans Cybern. https://doi.org/10.1109/TCYB.2021.3064821

    Article  Google Scholar 

  48. Huang GB, Zhou H, Ding X, Zhang R (2011) Extreme learning machine for regression and multiclass classification. IEEE Trans Sys, Man Cybern, Part B (Cybernetics) 42(2):513–529

    Article  Google Scholar 

  49. Huang GB, Zhu QY, Siew CK (2004) Extreme learning machine: a new learning scheme of feedforward neural networks. In 2004 IEEE International joint conference on neural networks (IEEE Cat. No. 04CH37541) (Vol. 2, pp. 985-990). IEEE

  50. Liu Z, Cao W, Gao Z, Bian J, Chen H, Chang Y, Liu T-Y (2020) Self-paced ensemble for highly imbalanced massive data classification. In 2020 IEEE 36th International conference on data engineering (ICDE) (pp. 841-852), April. IEEE

  51. Jinyan L, Simon F, Wong Raymond K, Chu Victor W (2018) Adaptive multi-objective swarm fusion for imbalanced data classification. Info Fus 39:1–24

    Article  Google Scholar 

  52. Asuncion A, Newman D (2007) UCI machine learning repository

  53. Alcalá-Fdez J, Fernández A, Luengo J, Derrac J, García S, Sánchez L, Herrera F (2011) Keel data-mining software tool: data set repository, integration of algorithms and experimental analysis framework. J Mult-Val Log Soft Comput, 17

  54. Menzies T, Greenwald J, Frank A (2006) Data mining static code attributes to learn defect predictors. IEEE Trans Softw Eng 33(1):2–13

    Article  Google Scholar 

  55. Song Q, Jia Z, Shepperd M, Ying S, Liu J (2010) A general software defect-proneness prediction framework. IEEE Trans Softw Eng 37(3):356–370

    Article  Google Scholar 

  56. Gong LN, Jiang SJ, Jiang L (2019) Research progress of software defect prediction. Ruan Jian Xue Bao/J Softw 30(10):3090–3114

    Google Scholar 

  57. Song Q, Guo Y, Shepperd M (2018) A comprehensive investigation of the role of imbalanced learning for software defect prediction. IEEE Trans Softw Eng 45(12):1253–1269

    Article  Google Scholar 

  58. Martin Shepperd et al (2013) Data quality: some comments on the NASA software defect datasets. IEEE Trans Softw Eng 39(9):1208–1215

    Article  Google Scholar 

  59. Wang S, Yao X (2013) Using class imbalance learning for software defect prediction. IEEE Trans Reliab 62(2):434–443

    Article  Google Scholar 

Download references

Funding

This article was funded by Key-Area Research and Development Program of Guangdong Province under the Grant No. (2019B010153002) and National Natural Science Foundation of China under the Grant No. (62106224) and Marine Economy Development (Six Marine Industries) Special Foundation of Department of Natural Resources of Guangdong Province under Grant GDNRC [2020]056.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Kaixiang Yang.

Ethics declarations

Conflict of interest

The authors declare that they have no conflict of interest.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Chen, W., Yang, K., Zhang, W. et al. Double-kernelized weighted broad learning system for imbalanced data. Neural Comput & Applic 34, 19923–19936 (2022). https://doi.org/10.1007/s00521-022-07534-5

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s00521-022-07534-5

Keywords

Navigation