Nothing Special   »   [go: up one dir, main page]

skip to main content
10.1145/3511808.3557474acmconferencesArticle/Chapter ViewAbstractPublication PagescikmConference Proceedingsconference-collections
research-article
Public Access

Towards Automated Imbalanced Learning with Deep Hierarchical Reinforcement Learning

Published: 17 October 2022 Publication History

Abstract

Imbalanced learning is a fundamental challenge in data mining, where there is a disproportionate ratio of training samples in each class. Over-sampling is an effective technique to tackle imbalanced learning through generating synthetic samples for the minority class. While numerous over-sampling algorithms have been proposed, they heavily rely on heuristics, which could be sub-optimal since we may need different sampling strategies for different datasets and base classifiers, and they cannot directly optimize the performance metric. Motivated by this, we investigate developing a learning-based over-sampling algorithm to optimize the classification performance, which is a challenging task because of the huge and hierarchical decision space. At the high level, we need to decide how many synthetic samples to generate. At the low level, we need to determine where the synthetic samples should be located, which depends on the high-level decision since the optimal locations of the samples may differ for different numbers of samples. To address the challenges, we propose AutoSMOTE, an automated over-sampling algorithm that can jointly optimize different levels of decisions. Motivated by the success of SMOTE and its extensions, we formulate the generation process as a Markov decision process (MDP) consisting of three levels of policies to generate synthetic samples within the SMOTE search space. Then we leverage deep hierarchical reinforcement learning to optimize the performance metric on the validation data. Extensive experiments on six real-world datasets demonstrate that AutoSMOTE significantly outperforms the state-of-the-art resampling algorithms. The code is at https://github.com/daochenzha/autosmote

Supplementary Material

MP4 File (CIKM22-fp0772.mp4)
Presentation video

References

[1]
Sabri Boughorbel, Fethi Jarray, and Mohammed El-Anbari. 2017. Optimal classifier for imbalanced data using Matthews Correlation Coefficient metric. PloS one, Vol. 12, 6 (2017), e0177678.
[2]
Mateusz Buda, Atsuto Maki, and Maciej A Mazurowski. 2018. A systematic study of the class imbalance problem in convolutional neural networks. Neural Networks, Vol. 106 (2018), 249--259.
[3]
Zixin Cai, Xinyue Wang, Mingjie Zhou, Jian Xu, and Liping Jing. 2019. Supervised class distribution learning for GANs-based imbalanced classification. In ICDM.
[4]
Nitesh V Chawla, Kevin W Bowyer, Lawrence O Hall, and W Philip Kegelmeyer. 2002. SMOTE: synthetic minority over-sampling technique. Journal of artificial intelligence research, Vol. 16 (2002), 321--357.
[5]
Nitesh V Chawla, Nathalie Japkowicz, and Aleksander Kotcz. 2004. Special issue on learning from imbalanced data sets. ACM SIGKDD explorations newsletter, Vol. 6, 1 (2004), 1--6.
[6]
Thomas Elsken, Jan Hendrik Metzen, and Frank Hutter. 2019. Neural architecture search: A survey. The Journal of Machine Learning Research, Vol. 20, 1 (2019), 1997--2017.
[7]
Lasse Espeholt, Hubert Soyer, Remi Munos, Karen Simonyan, Vlad Mnih, Tom Ward, Yotam Doron, Vlad Firoiu, Tim Harley, Iain Dunning, et al. 2018. Impala: Scalable distributed deep-rl with importance weighted actor-learner architectures. In ICML.
[8]
Hui Han, Wen-Yuan Wang, and Bing-Huan Mao. 2005. Borderline-SMOTE: a new over-sampling method in imbalanced data sets learning. In ICIC.
[9]
Songqiao Han, Xiyang Hu, Hailiang Huang, Mingqi Jiang, and Yue Zhao. 2022. ADBench: Anomaly Detection Benchmark. arXiv preprint arXiv:2206.09426 (2022).
[10]
Haibo He, Yang Bai, Edwardo A Garcia, and Shutao Li. 2008. ADASYN: Adaptive synthetic sampling approach for imbalanced learning. In IJCNN.
[11]
Yuval Heffetz, Roman Vainshtein, Gilad Katz, and Lior Rokach. 2020. Deepline: Automl tool for pipelines generation using deep reinforcement learning and hierarchical actions filtering. In KDD.
[12]
Justin M Johnson and Taghi M Khoshgoftaar. 2019. Survey on deep learning with class imbalance. Journal of Big Data, Vol. 6, 1 (2019), 1--54.
[13]
Patrick Koch, Oleg Golovidov, Steven Gardner, Brett Wujek, Joshua Griffin, and Yan Xu. 2018. Autotune: A derivative-free optimization framework for hyperparameter tuning. In KDD.
[14]
György Kovács. 2019a. An empirical comparison and evaluation of minority oversampling techniques on a large number of imbalanced datasets. Applied Soft Computing, Vol. 83 (2019), 105662.
[15]
György Kovács. 2019b. Smote-variants: A python implementation of 85 minority oversampling techniques. Neurocomputing, Vol. 366 (2019), 352--354.
[16]
Bartosz Krawczyk. 2016. Learning from imbalanced data: open challenges and future directions. Progress in Artificial Intelligence, Vol. 5, 4 (2016), 221--232.
[17]
Kwei-Herng Lai, Daochen Zha, Guanchu Wang, Junjie Xu, Yue Zhao, Devesh Kumar, Yile Chen, Purav Zumkhawaka, Minyang Wan, Diego Martinez, et al. 2021a. TODS: An Automated Time Series Outlier Detection System. In AAAI.
[18]
Kwei-Herng Lai, Daochen Zha, Junjie Xu, Yue Zhao, Guanchu Wang, and Xia Hu. 2021b. Revisiting time series outlier detection: Definitions and benchmarks. In NeurIPS.
[19]
Mingchen Li, Xuechen Zhang, Christos Thrampoulidis, Jiasi Chen, and Samet Oymak. 2021c. AutoBalance: Optimized Loss Functions for Imbalanced Data. NeurIPS (2021).
[20]
Ting Li, Junbo Zhang, Kainan Bao, Yuxuan Liang, Yexin Li, and Yu Zheng. 2020b. Autost: Efficient neural architecture search for spatio-temporal prediction. In KDD.
[21]
Yuening Li, Zhengzhang Chen, Daochen Zha, Kaixiong Zhou, Haifeng Jin, Haifeng Chen, and Xia Hu. 2021a. Automated Anomaly Detection via Curiosity-Guided Search and Self-Imitation Learning. IEEE Transactions on Neural Networks and Learning Systems (2021).
[22]
Yuening Li, Zhengzhang Chen, Daochen Zha, Kaixiong Zhou, Haifeng Jin, Haifeng Chen, and Xia Hu. 2021b. Autood: Neural architecture search for outlier detection. In ICDE.
[23]
Yuening Li, Daochen Zha, Praveen Venugopal, Na Zou, and Xia Hu. 2020a. Pyodds: An end-to-end outlier detection system with automated machine learning. In WWW.
[24]
Kunpeng Liu, Yanjie Fu, Pengfei Wang, Le Wu, Rui Bo, and Xiaolin Li. 2019. Automating feature subspace exploration via multi-agent reinforcement learning. In KDD.
[25]
Xu-Ying Liu and Zhi-Hua Zhou. 2006. The influence of class imbalance on cost-sensitive learning: An empirical study. In ICDM.
[26]
Zhining Liu, Pengfei Wei, Jing Jiang, Wei Cao, Jiang Bian, and Yi Chang. 2020. MESA: Boost Ensemble Imbalanced Learning with MEta-SAmpler. In NeurIPS.
[27]
Volodymyr Mnih, Koray Kavukcuoglu, David Silver, Alex Graves, Ioannis Antonoglou, Daan Wierstra, and Martin Riedmiller. 2013. Playing atari with deep reinforcement learning. arXiv preprint arXiv:1312.5602 (2013).
[28]
Ajinkya More. 2016. Survey of resampling techniques for improving classification performance in unbalanced datasets. arXiv preprint arXiv:1608.06048 (2016).
[29]
Hien M Nguyen, Eric W Cooper, and Katsuari Kamei. 2011. Borderline over-sampling for imbalanced data classification. International Journal of Knowledge Engineering and Soft Data Paradigms, Vol. 3, 1 (2011), 4--21.
[30]
Shubham Pateria, Budhitama Subagdja, Ah-hwee Tan, and Chai Quek. 2021. Hierarchical Reinforcement Learning: A Comprehensive Survey. ACM Computing Surveys (CSUR), Vol. 54, 5 (2021), 1--35.
[31]
Neelam Rout, Debahuti Mishra, and Manas Kumar Mallick. 2018. Handling imbalanced data: a survey. In ASISA.
[32]
B Santoso, H Wijayanto, KA Notodiputro, and B Sartono. 2017. Synthetic over sampling methods for handling class imbalanced problems: A review. In IOP conference series: earth and environmental science, Vol. 58. IOP Publishing, 012031.
[33]
Kai Shu, Amy Sliva, Suhang Wang, Jiliang Tang, and Huan Liu. 2017. Fake news detection on social media: A data mining perspective. ACM SIGKDD explorations newsletter, Vol. 19, 1 (2017), 22--36.
[34]
Wacharasak Siriseriwan and Krung Sinapiromsaran. 2017. Adaptive neighbor synthetic minority oversampling technique under 1NN outcast handling. Songklanakarin J. Sci. Technol, Vol. 39, 5 (2017), 565--576.
[35]
Richard S Sutton and Andrew G Barto. 2018. Reinforcement learning: An introduction. MIT press.
[36]
Joaquin Vanschoren, Jan N Van Rijn, Bernd Bischl, and Luis Torgo. 2014. OpenML: networked science in machine learning. ACM SIGKDD Explorations Newsletter, Vol. 15, 2 (2014), 49--60.
[37]
Jing Wang and Min-Ling Zhang. 2018. Towards mitigating the class-imbalance problem for partial label learning. In KDD.
[38]
Wentao Wang, Suhang Wang, Wenqi Fan, Zitao Liu, and Jiliang Tang. 2020. Global-and-local aware data generation for the class imbalance problem. In SDM.
[39]
Yicheng Wang, Xiaotian Han, Chia-Yuan Chang, Daochen Zha, Ulisses Braga-Neto, and Xia Hu. 2022. Auto-PINN: Understanding and Optimizing Physics-Informed Neural Architecture. arXiv preprint arXiv:2205.13748 (2022).
[40]
Wikipedia. 2022. Wilcoxon signed-rank test -- Wikipedia, The Free Encyclopedia. http://en.wikipedia.org/w/index.php?title=Wilcoxon%20signed-rank%20test&oldid=1084875027. [Online; accessed 19-May-2022].
[41]
Lei Xu, Maria Skoularidou, Alfredo Cuesta-Infante, and Kalyan Veeramachaneni. 2019. Modeling Tabular data using Conditional GAN. In NeurIPS.
[42]
Chengrun Yang, Yuji Akimoto, Dae Won Kim, and Madeleine Udell. 2019. OBOE: Collaborative filtering for AutoML model selection. In KDD.
[43]
Chengrun Yang, Jicong Fan, Ziyang Wu, and Madeleine Udell. 2020. Automl pipeline selection: Efficiently navigating the combinatorial space. In KDD.
[44]
Show-Jane Yen and Yue-Shi Lee. 2006. Under-sampling approaches for improving prediction of the minority class in an imbalanced dataset. In Intelligent Control and Automation. Springer, 731--740.
[45]
Tong Yu and Hong Zhu. 2020. Hyper-parameter optimization: A review of algorithms and applications. arXiv preprint arXiv:2003.05689 (2020).
[46]
Daochen Zha, Louis Feng, Bhargav Bhushanam, Dhruv Choudhary, Jade Nie, Yuandong Tian, Jay Chae, Yinbin Ma, Arun Kejariwal, and Xia Hu. 2022a. AutoShard: Automated Embedding Table Sharding for Recommender Systems. In KDD.
[47]
Daochen Zha, Kwei-Herng Lai, Songyi Huang, Yuanpu Cao, Keerthana Reddy, Juan Vargas, Alex Nguyen, Ruzhe Wei, Junyu Guo, and Xia Hu. 2021a. RLCard: a platform for reinforcement learning in card games. In IJCAI.
[48]
Daochen Zha, Kwei-Herng Lai, Mingyang Wan, and Xia Hu. 2020. Meta-AAD: Active anomaly detection with deep reinforcement learning. In ICDM.
[49]
Daochen Zha, Kwei-Herng Lai, Kaixiong Zhou, and Xia Hu. 2019. Experience Replay Optimization. In IJCAI.
[50]
Daochen Zha, Kwei-Herng Lai, Kaixiong Zhou, and Xia Hu. 2021b. Simplifying deep reinforcement learning via self-supervision. arXiv preprint arXiv:2106.05526 (2021).
[51]
Daochen Zha, Wenye Ma, Lei Yuan, Xia Hu, and Ji Liu. 2021c. Rank the Episodes: A Simple Approach for Exploration in Procedurally-Generated Environments. In ICLR.
[52]
Daochen Zha, Zaid Pervaiz Bhat, Yi-Wei Chen, Yicheng Wang, Sirui Ding, Anmoll Kumar Jain, Mohammad Qazim Bhat, Kwei-Herng Lai, Jiaben Chen, et al. 2022b. AutoVideo: An Automated Video Action Recognition System. In IJCAI.
[53]
Daochen Zha, Jingru Xie, Wenye Ma, Sheng Zhang, Xiangru Lian, Xia Hu, and Ji Liu. 2021d. DouZero: Mastering DouDizhu with Self-Play Deep Reinforcement Learning. In ICML.
[54]
Tianxiang Zhao, Xiang Zhang, and Suhang Wang. 2021c. Graphsmote: Imbalanced node classification on graphs with graph neural networks. In WSDM.
[55]
Xiangyu Zhao, Haochen Liu, Wenqi Fan, Hui Liu, Jiliang Tang, Chong Wang, Ming Chen, Xudong Zheng, Xiaobing Liu, and Xiwang Yang. 2021a. Autoemb: Automated embedding dimensionality search in streaming recommendations. In ICDM.
[56]
Yue Zhao, Ryan Rossi, and Leman Akoglu. 2021b. Automatic unsupervised outlier model selection. NeurIPS (2021).

Cited By

View all
  • (2024)How Automated Machine Learning Can Improve BusinessApplied Sciences10.3390/app1419874914:19(8749)Online publication date: 27-Sep-2024
  • (2024)Energy-efficient access point clustering and power allocation in cell-free massive MIMO networks: a hierarchical deep reinforcement learning approachEURASIP Journal on Advances in Signal Processing10.1186/s13634-024-01111-92024:1Online publication date: 26-Jan-2024
  • (2023)Bias in Reinforcement Learning: A Review in Healthcare ApplicationsACM Computing Surveys10.1145/360950256:2(1-17)Online publication date: 15-Sep-2023
  • Show More Cited By

Index Terms

  1. Towards Automated Imbalanced Learning with Deep Hierarchical Reinforcement Learning

      Recommendations

      Comments

      Please enable JavaScript to view thecomments powered by Disqus.

      Information & Contributors

      Information

      Published In

      cover image ACM Conferences
      CIKM '22: Proceedings of the 31st ACM International Conference on Information & Knowledge Management
      October 2022
      5274 pages
      ISBN:9781450392365
      DOI:10.1145/3511808
      • General Chairs:
      • Mohammad Al Hasan,
      • Li Xiong
      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

      Sponsors

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      Published: 17 October 2022

      Permissions

      Request permissions for this article.

      Check for updates

      Author Tags

      1. automated machine learning
      2. classification
      3. imbalanced learning
      4. reinforcement learning

      Qualifiers

      • Research-article

      Funding Sources

      Conference

      CIKM '22
      Sponsor:

      Acceptance Rates

      CIKM '22 Paper Acceptance Rate 621 of 2,257 submissions, 28%;
      Overall Acceptance Rate 1,861 of 8,427 submissions, 22%

      Upcoming Conference

      CIKM '25

      Contributors

      Other Metrics

      Bibliometrics & Citations

      Bibliometrics

      Article Metrics

      • Downloads (Last 12 months)194
      • Downloads (Last 6 weeks)16
      Reflects downloads up to 14 Nov 2024

      Other Metrics

      Citations

      Cited By

      View all
      • (2024)How Automated Machine Learning Can Improve BusinessApplied Sciences10.3390/app1419874914:19(8749)Online publication date: 27-Sep-2024
      • (2024)Energy-efficient access point clustering and power allocation in cell-free massive MIMO networks: a hierarchical deep reinforcement learning approachEURASIP Journal on Advances in Signal Processing10.1186/s13634-024-01111-92024:1Online publication date: 26-Jan-2024
      • (2023)Bias in Reinforcement Learning: A Review in Healthcare ApplicationsACM Computing Surveys10.1145/360950256:2(1-17)Online publication date: 15-Sep-2023
      • (2023)Tackling Diverse Minorities in Imbalanced ClassificationProceedings of the 32nd ACM International Conference on Information and Knowledge Management10.1145/3583780.3615071(1178-1187)Online publication date: 21-Oct-2023
      • (2022)DreamShardProceedings of the 36th International Conference on Neural Information Processing Systems10.5555/3600270.3601375(15190-15203)Online publication date: 28-Nov-2022

      View Options

      View options

      PDF

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader

      Get Access

      Login options

      Media

      Figures

      Other

      Tables

      Share

      Share

      Share this Publication link

      Share on social media