Abstract
Data augmentation has been an essential technique to increase the amount and diversity of datasets, thus improving deep learning models. To pursue further performance, several automated data augmentation approaches have recently been proposed to find data augmentation policies automatically. However, there are still some key issues that deserve further exploration, i.e., a precise policy search space definition, the instructive policy evaluation method, and the low computational cost of policy search. In this paper, we propose a novel method named BO-Aug that attempts to solve the above issues. Empirical verification on three widely used image classification datasets shows that the proposed method can achieve state-of-the-art or comparable performance compared with advanced automated data augmentation methods, with a relatively low cost. Our code is available at https://github.com/Zhangcx19/BO-Aug.
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.References
Cubuk ED, Zoph B, Mane D, Vasudevan V, Le QV (2019) Autoaugment: learning augmentation policies from data. In: Proceedings of the 32nd IEEE conference on computer vision and pattern recognition, pp 113–123
Ho D, Liang E, Chen X, Stoica I, Abbeel P (2019) Population based augmentation: efficient learning of augmentation policy schedules. In: Proceedings of the 36th international conference on machine learning, pp 2731–2741
Lim S, Kim I, Kim T, Kim C, Kim S (2019) Fast autoaugment. In: Proceedings of the 33rd advances in neural information processing systems, pp 6665–6675
Takase T, Karakida R, Asoh H (2021) Self-paced data augmentation for training neural networks. Neurocomputing 442:296–306
Bandara K, Hewamalage H, Liu Y-H, Kang Y, Bergmeir C (2021) Improving the accuracy of global forecasting models using time series data augmentation. Pattern Recogn 120:108148
Wang Y, Wei X, Tang X, Shen H, Ding L (2020) Cnn tracking based on data augmentation. Knowl-Based Syst 194:105594
Leng Y, Zhao W, Lin C, Sun C, Wang R, Yuan Q, Li D (2020) Lda-based data augmentation algorithm for acoustic scene classification. Knowl-Based Syst 195:105600
Yun S, Han D, Oh SJ, Chun S, Choe J, Yoo Y (2019) Cutmix: regularization strategy to train strong classifiers with localizable features. In: Proceedings of the 17th international conference on computer vision, pp 6023–6032
Tolstikhin IO, Houlsby N, Kolesnikov A, Beyer L, Zhai X, Unterthiner T, Yung J, Steiner A, Keysers D, Uszkoreit J et al (2021) Mlp-mixer: an all-mlp architecture for vision. Adv Neural Inf Process Syst, vol 34
Sohn K, Berthelot D, Carlini N, Zhang Z, Zhang H, Raffel C, Cubuk ED, Kurakin A, Li C (2020) Fixmatch: simplifying semi-supervised learning with consistency and confidence. In: Proceedings of the 34th advances in neural information processing systems, pp 596–608
Zoph B, Ghiasi G, Lin T, Cui Y, Liu H, Cubuk ED, Le Q (2020) Rethinking pre-training and self-training. In: Proceedings of the 34th advances in neural information processing systems, pp 3833–3845
Xie Q, Dai Z, Hovy EH, Luong T, Le Q (2020) Unsupervised data augmentation for consistency training. In: Proceedings of the 34th advances in neural information processing systems, pp 6256–6268
Caron M, Misra I, Mairal J, Goyal P, Bojanowski P, Joulin A (2020) Unsupervised learning of visual features by contrasting cluster assignments. In: Proceedings of the 34th advances in neural information processing systems, pp 9912–9924
He K, Fan H, Wu Y, Xie S, Girshick RB (2020) Momentum contrast for unsupervised visual representation learning. In: Proceedings of the 34th advances in neural information processing systems, pp 9726–9735
Misra I, Maaten LVD (2020) Self-supervised learning of pretext-invariant representations. In: Proceedings of the 33rd IEEE conference on computer vision and pattern recognition, pp 6706–6716
Chen T, Kornblith S, Swersky K, Norouzi M, Hinton GE (2020) Big self-supervised models are strong semi-supervised learners. In: Proceedings of the 34th advances in neural information processing systems, pp 22243–22255
Chen T, Kornblith S, Norouzi M, Hinton GE (2020) A simple framework for contrastive learning of visual representations. In: Proceedings of the 37th international conference on machine learning, pp 1597–1607
Hu J, Shen L, Sun G (2018) Squeeze-and-excitation networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 7132–7141
Tan M, Le Q (2019) Efficientnet: rethinking model scaling for convolutional neural networks. In: International conference on machine learning, pp 6105–6114
DeVries T, Taylor GW (2017) Improved regularization of convolutional neural networks with cutout. arXiv:1708.04552
Zhong Z, Zheng L, Kang G, Li S, Yang Y (2020) Random erasing data augmentation. In: Proceedings of the AAAI conference on artificial intelligence, vol 34, pp 13001–13008
Zhang H, Cisse M, Dauphin YN, Lopez-paz D (2018) Mixup: beyond empirical risk minimization. In: International conference on learning representations
Chinbat V, Bae S-H (2022) Ga3n: generative adversarial autoaugment network. Pattern Recogn 127:108637
dos Santos Tanaka FHK, Aranha C (2019) Data augmentation using gans. Proc Mach Learn Res 1:16
Songyan Liu J-GH, Haiyun Guo X u, Zhao EA (2020) A novel data augmentation scheme for pedestrian detection with attribute preserving gan. Neurocomputing 401:123–132
DeVries T, Taylor GW (2017) Dataset augmentation in feature space. In: Proceedings of the 34th international conference on machine learning, workshop track
Wang Y, Pan X, Song S, Zhang H, Huang G, Wu C (2019) Implicit semantic data augmentation for deep networks. Adv Neural Inf Process Syst, vol 32
Pham H, Guan M, Zoph B, Le Q, Dean J (2018) Efficient neural architecture search via parameters sharing. In: International conference on machine learning, pp 4095–4104
Ding Z, Chen Y, Li N, Zhao D, Sun Z, Chen CP (2021) Bnas: efficient neural architecture search using broad scalable architecture. IEEE Trans Neural Netw Learn Syst
Wei C, Niu C, Tang Y, Wang Y, Hu H, Liang J (2022) Npenas: neural predictor guided evolution for neural architecture search. IEEE Trans Neural Netw Learn Syst
Chen Y, Gao R, Liu F, Zhao D (2021) Modulenet: knowledge-inherited neural architecture search. IEEE Trans Cybern
Cubuk ED, Zoph B, Shlens J, Le QV (2020) Randaugment: practical automated data augmentation with a reduced search space. In: Proceedings of the 33rd IEEE conference on computer vision and pattern recognition, pp 3008–3017
Naghizadeh A, Abavisani M, Metaxas DN (2020) Greedy autoaugment. Pattern Recogn Lett 138:624–630
Tian K, Lin C, Sun M, Zhou L, Yan J, Ouyang W (2020) Improving auto-augment via augmentation-wise weight sharing. Adv Neural Inf Process Syst 33:19088–19098
Cui J, Yang B (2017) Survey on bayesian optimization methodology and applications. In: Journal of software, pp 176–198
Lindauer M, Eggensperger K, Feurer M, Biedenkapp A, Deng D, Benjamins C, Ruhkopf T, Sass R, Hutter F (2022) Smac3: a versatile bayesian optimization package for hyperparameter optimization. J Mach Learn Res 23(54):1–9
Imani M, Ghoreishi SF (2021) Graph-based bayesian optimization for large-scale objective-based experimental design. IEEE Trans Neural Netw Learn Syst
Du L, Gao R, Suganthan PN, Wang DZ (2022) Bayesian optimization based dynamic ensemble for time series forecasting. Inf Sci 591:155–175
Turner R, Eriksson D, McCourt M, Kiili J, Laaksonen E, Xu Z, Guyon I (2021) Bayesian optimization is superior to random search for machine learning hyperparameter tuning: analysis of the black-box optimization challenge 2020. In: NeurIPS 2020 competition and demonstration track, pp 3–26
Williams CK, Rasmussen CE (2006) Gaussian Processes for Machine Learning. MIT Press Cambridge, MA
Krizhevsky A, Hinton G (2009) Learning multiple layers of features from tiny images. Technical report, Citeseer
Netzer Y, Wang T, Coates A, Bissacco A, Wu B, Ng AY (2011) Reading digits in natural images with unsupervised feature learning. In: Proceedings of the 28th neural information processing systems workshop on deep learning and unsupervised feature learning
Deng J, Dong W, Socher R, Li L-J, Li K, Fei-Fei L (2009) Imagenet: a large-scale hierarchical image database. In: IEEE conference on computer vision and pattern recognition, pp 248–255
Zagoruyko S, Komodakis N (2016) Wide residual networks. Proceedings of the 27th British machine vision conference, pp 8701–8712
Gastaldi X (2017) Shake-shake regularization. arXiv:1705.07485
Acknowledgements
This work was supported by the National Key R&D Program of China under Grant Nos. 2021ZD0112501 and 2021ZD0112502; the National Natural Science Foundation of China under Grant Nos. 62172185 and 61876069; Jilin Province Key Scientific and Technological Research and Development Project under Grant Nos. 20180201067GX and 20180201044GX; Jilin Province Natural Science Foundation under Grant No. 20200201036JC; and China Postdoctoral Science Foundation funded project under Grant No.2021M701388.
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher’s note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Appendices
Appendix A: Image transformation operations used for DA policies search
All available image transformation functions during the search process are listed below. These functions accept an image and corresponding operation parameters as input, and output a transformed image. The range of magnitudes for each operation is shown in the third column. Some transformations do not use magnitude information (e.g., Invert and Equalize).
Appendix B: Mapping relationship between operation type value of policy vector and operation types
For each policy, we regard its first dimension as the two DA operation types. Considering that there are 14 DA operations in total, we use a number between 0 and 196 to represent two operations. Here we list the mapping relationship of the value and corresponding operation types.
Appendix C: Hyperparameter configuration in the experiments
Once BO-Aug finds the optimal DA policies, we will validate the policies’ performance on different target models based on several datasets. Here we give the learning rate and weight decay values during target model training.
Appendix D: Policies found on reduced CIFAR-10
Appendix E: Policies found on reduced SVHN
Appendix F: Policies found on reduced TinyImagenet
Rights and permissions
About this article
Cite this article
Zhang, C., Li, X., Zhang, Z. et al. BO-Aug: learning data augmentation policies via Bayesian optimization. Appl Intell 53, 8978–8993 (2023). https://doi.org/10.1007/s10489-022-03790-z
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10489-022-03790-z