Nothing Special   »   [go: up one dir, main page]

skip to main content
10.1145/3394171.3413978acmconferencesArticle/Chapter ViewAbstractPublication PagesmmConference Proceedingsconference-collections
research-article

CRSSC: Salvage Reusable Samples from Noisy Data for Robust Learning

Published: 12 October 2020 Publication History

Abstract

Due to the existence of label noise in web images and the high memorization capacity of deep neural networks, training deep fine-grained (FG) models directly through web images tends to have an inferior recognition ability. In the literature, to alleviate this issue, loss correction methods try to estimate the noise transition matrix, but the inevitable false correction would cause severe accumulated errors. Sample selection methods identify clean ("easy") samples based on the fact that small losses can alleviate the accumulated errors. However, "hard" and mislabeled examples that can both boost the robustness of FG models are also dropped. To this end, we propose a certainty-based reusable sample selection and correction approach, termed as CRSSC, for coping with label noise in training deep FG models with web images. Our key idea is to additionally identify and correct reusable samples, and then leverage them together with clean examples to update the networks. We demonstrate the superiority of the proposed approach from both theoretical and experimental perspectives.

Supplementary Material

MP4 File (3394171.3413978.mp4)
Due to the existence of label noise in web images and the high memorization capacity of deep neural networks, training deep fine-grained (FG) models directly through web images tends to have an inferior recognition ability. In the literature, to alleviate this issue, loss correction methods try to estimate the noise transition matrix, but the inevitable false correction could accumulate errors. Sample selection methods identify clean samples based on a low-loss criterion. However, ?hard? and mislabeled examples that can both boost the robustness of FG models are also dropped. To this end, we propose a certainty-based reusable sample selection and correction approach, termed as CRSSC, for coping with label noise in training deep FG models with web images. Our key idea is to additionally identify and correct reusable samples, and then leverage them together with clean examples to update the networks. We demonstrate the superiority of the proposed approach from both theoretical and experimental perspectives.

References

[1]
Devansh Arpit, Stanisław Jastrzke bski, Nicolas Ballas, David Krueger, Emmanuel Bengio, Maxinder S Kanwal, Tegan Maharaj, Asja Fischer, Aaron Courville, Yoshua Bengio, et al. 2017. A closer look at memorization in deep networks. In International Conference on Machine Learning. 233--242.
[2]
Steve Branson, Grant Van Horn, Serge Belongie, and Pietro Perona. 2014. Bird species categorization using pose normalized deep convolutional nets. In British Machine Vision Conference.
[3]
Haw-Shiuan Chang, Erik Learned-Miller, and Andrew McCallum. 2017a. Active bias: Training more accurate neural networks by emphasizing high variance samples. In Advances in Neural Information Processing Systems. 1002--1012.
[4]
Haw-Shiuan Chang, Erik Learned-Miller, and Andrew McCallum. 2017b. Active bias: Training more accurate neural networks by emphasizing high variance samples. In Advances in Neural Information Processing Systems. 1002--1012.
[5]
Tao Chen, Jian Zhang, Guo-Sen Xie, Yazhou Yao, Xiaoshui Huang, and Zhenmin Tang. 2020. Classification Constrained Discriminator For Domain Adaptive Semantic Segmentation. In IEEE International Conference on Multimedia and Expo. 1--6.
[6]
Yue Chen, Yalong Bai, Wei Zhang, and Tao Mei. 2019. Destruction and construction learning for fine-grained image recognition. In IEEE Conference on Computer Vision and Pattern Recognition. 5157--5166.
[7]
Yin Cui, Yang Song, Chen Sun, Andrew Howard, and Serge Belongie. 2018. Large scale fine-grained categorization and domain-specific transfer learning. In IEEE Conference on Computer Vision and Pattern Recognition. 4109--4118.
[8]
Yin Cui, Feng Zhou, Yuanqing Lin, and Serge Belongie. 2016. Fine-grained categorization and dataset bootstrapping using deep metric learning with humans in the loop. In IEEE Conference on Computer Vision and Pattern Recognition. 1153--1162.
[9]
Yin Cui, Feng Zhou, Jiang Wang, Xiao Liu, Yuanqing Lin, and Serge Belongie. 2017. Kernel pooling for convolutional neural networks. In IEEE Conference on Computer Vision and Pattern Recognition. 2921--2930.
[10]
Abhimanyu Dubey, Otkrist Gupta, Ramesh Raskar, and Nikhil Naik. 2018. Maximum-entropy fine grained classification. In Advances in Neural Information Processing Systems. 637--647.
[11]
Yang Gao, Oscar Beijbom, Ning Zhang, and Trevor Darrell. 2016. Compact bilinear pooling. In IEEE Conference on Computer Vision and Pattern Recognition. 317--326.
[12]
Weifeng Ge, Xiangru Lin, and Yizhou Yu. 2019. Weakly supervised complementary parts models for fine-grained image classification from the bottom up. In IEEE Conference on Computer Vision and Pattern Recognition. 3034--3043.
[13]
Jacob Goldberger and Ehud Ben-Reuven. 2017. Training deep neural-networks using a noise adaptation layer. In International Conference on Learning Representations.
[14]
Bo Han, Quanming Yao, Xingrui Yu, Gang Niu, Miao Xu, Weihua Hu, Ivor Tsang, and Masashi Sugiyama. 2018. Co-teaching: Robust training of deep neural networks with extremely noisy labels. In Advances in Neural Information Processing Systems. 8527--8537.
[15]
Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. 2016. Deep residual learning for image recognition. In IEEE Conference on Computer Vision and Pattern Recognition. 770--778.
[16]
Xian-sheng Hua, Fumin Shen, Jian Zhang, and Zhenmin Tang. 2016. A domain robust approach for image dataset construction. In ACM international conference on Multimedia. 212--216.
[17]
Shaoli Huang, Zhe Xu, Dacheng Tao, and Ya Zhang. 2016. Part-stacked cnn for fine-grained visual categorization. In IEEE Conference on Computer Vision and Pattern Recognition. 1173--1182.
[18]
Lu Jiang, Zhengyuan Zhou, Thomas Leung, Li-Jia Li, and Li Fei-Fei. 2017. Mentornet: Learning data-driven curriculum for very deep neural networks on corrupted labels. In International Conference on Machine Learning.
[19]
Shu Kong and Charless Fowlkes. 2017. Low-rank bilinear pooling for fine-grained classification. In IEEE Conference on Computer Vision and Pattern Recognition. 365--374.
[20]
Jonathan Krause, Michael Stark, Jia Deng, and Li Fei-Fei. 2013. 3d object representations for fine-grained categorization. In CVPRW. 554--561.
[21]
Alex Krizhevsky and Geoffrey Hinton. 2009. Learning multiple layers of features from tiny images. Technical report, University of Toronto, Vol. 1, 4 (2009), 7.
[22]
M Pawan Kumar, Benjamin Packer, and Daphne Koller. 2010. Self-paced learning for latent variable models. In Advances in Neural Information Processing Systems. 1189--1197.
[23]
Michael Lam, Behrooz Mahasseni, and Sinisa Todorovic. 2017. Fine-grained recognition as hsnet search for informative image parts. In IEEE Conference on Computer Vision and Pattern Recognition. 2520--2529.
[24]
Peihua Li, Jiangtao Xie, Qilong Wang, and Zilin Gao. 2018. Towards faster training of global covariance pooling networks by iterative matrix square root normalization. In IEEE Conference on Computer Vision and Pattern Recognition. 947--955.
[25]
Tsung-Yi Lin, Priya Goyal, Ross Girshick, Kaiming He, and Piotr Dollár. 2017a. Focal loss for dense object detection. In IEEE International Conference on Computer Vision. 2980--2988.
[26]
Tsung-Yu Lin and Subhransu Maji. 2017. Improved bilinear pooling with cnns. In British Machine Vision Conference. 117.1--117.12.
[27]
Tsung-Yu Lin, Aruni RoyChowdhury, and Subhransu Maji. 2017b. Bilinear convolutional neural networks for fine-grained visual recognition., Vol. 40, 6 (2017), 1309--1322.
[28]
Jiarou Lu, Huafeng Liu, Yazhou Yao, Shuyin Tao, Zhenming Tang, and Jianfeng Lu. 2020. Hsi Road: A Hyper Spectral Image Dataset For Road Segmentation. In IEEE International Conference on Multimedia and Expo. 1--6.
[29]
Haonan Luo, Guosheng Lin, Zichuan Liu, Fayao Liu, Zhenmin Tang, and Yazhou Yao. 2019. Segeqa: Video segmentation based visual attention for embodied question answering. In IEEE International Conference on Computer Vision. 9667--9676.
[30]
Subhransu Maji, Esa Rahtu, Juho Kannala, Matthew Blaschko, and Andrea Vedaldi. 2013. Fine-grained visual classification of aircraft. arXiv:1306.5151 (2013).
[31]
Eran Malach and Shai Shalev-Shwartz. 2017. Decoupling "when to update" from "how to update". In Advances in Neural Information Processing Systems. 960--970.
[32]
Li Niu, Ashok Veeraraghavan, and Ashutosh Sabharwal. 2018. Webly supervised learning meets zero-shot learning: A hybrid approach for fine-grained classification. In IEEE Conference on Computer Vision and Pattern Recognition. 7171--7180.
[33]
Giorgio Patrini, Alessandro Rozza, Aditya Krishna Menon, Richard Nock, and Lizhen Qu. 2017. Making deep neural networks robust to label noise: A loss correction approach. In IEEE Conference on Computer Vision and Pattern Recognition. 1944--1952.
[34]
Scott Reed, Honglak Lee, Dragomir Anguelov, Christian Szegedy, Dumitru Erhan, and Andrew Rabinovich. 2014. Training deep neural networks on noisy labels with bootstrapping. In International Conference on Learning Representations. 1--11.
[35]
Mengye Ren, Wenyuan Zeng, Bin Yang, and Raquel Urtasun. 2018. Learning to reweight examples for robust deep learning. In International Conference on Machine Learning. 4334--4343.
[36]
Abhinav Shrivastava, Abhinav Gupta, and Ross Girshick. 2016. Training region-based object detectors with online hard example mining. In IEEE Conference on Computer Vision and Pattern Recognition. 761--769.
[37]
Xiangbo Shu, Jinhui Tang, Guojun Qi, Wei Liu, and Jian Yang. 2019. Hierarchical long short-term concurrent memory for human interaction recognition. IEEE transactions on pattern analysis and machine intelligence (2019).
[38]
Karen Simonyan and Andrew Zisserman. 2014. Very deep convolutional networks for large-scale image recognition. arXiv:1409.1556 (2014).
[39]
Hwanjun Song, Minseok Kim, and Jae-Gil Lee. 2019. SELFIE: Refurbishing unclean samples for robust deep learning. In International Conference on Machine Learning. 5907--5915.
[40]
Zeren Sun, Fumin Shen, Li Liu, and Limin et al. Wang. 2019. Dynamically visual disambiguation of keyword-based image search. International Joint Conference on Artificial Intelligence (2019), 996--1002.
[41]
Christian Szegedy, Vincent Vanhoucke, Sergey Ioffe, Jon Shlens, and Zbigniew Wojna. 2016. Rethinking the inception architecture for computer vision. In IEEE Conference on Computer Vision and Pattern Recognition. 2818--2826.
[42]
Jinhui Tang, Zechao Li, Hanjiang Lai, Liyan Zhang, Shuicheng Yan, et al. 2017. Personalized age progression with bi-level aging dictionary learning. IEEE transactions on pattern analysis and machine intelligence, Vol. 40, 4 (2017), 905--917.
[43]
Grant Van Horn, Oisin Mac Aodha, Yang Song, Yin Cui, Chen Sun, Alex Shepard, Hartwig Adam, Pietro Perona, and Serge Belongie. 2018. The inaturalist species classification and detection dataset. In IEEE Conference on Computer Vision and Pattern Recognition. 8769--8778.
[44]
Catherine Wah, Steve Branson, Peter Welinder, Pietro Perona, and Serge Belongie. 2011. The Caltech-UCSD Birds-200--2011 Dataset. CNS-TR-2011-001 (2011).
[45]
Xiu-Shen Wei, Chen-Wei Xie, Jianxin Wu, and Chunhua Shen. 2018. Mask-CNN: Localizing parts and selecting descriptors for fine-grained bird species categorization. Pattern Recognition, Vol. 76 (2018), 704--714.
[46]
Guo-Sen Xie, Li Liu, Xiaobo Jin, Fan Zhu, Zheng Zhang, Jie Qin, Yazhou Yao, and Ling Shao. 2019. Attentive region embedding network for zero-shot learning. In IEEE Conference on Computer Vision and Pattern Recognition. 9384--9393.
[47]
Guo-Sen Xie, Li Liu, Fan Zhu, Fang Zhao, Zheng Zhang, Yazhou Yao, Jie Qin, and Ling Shao. 2020. Region Graph Embedding Network for Zero-Shot Learning. In European Conference on Computer Vision.
[48]
Zhe Xu, Shaoli Huang, Ya Zhang, and Dacheng Tao. 2016. Webly-supervised fine-grained visual categorization via deep domain adaptation., Vol. 40, 5 (2016), 1100--1113.
[49]
Jufeng Yang, Xiaoxiao Sun, Yu-Kun Lai, Liang Zheng, and Ming-Ming Cheng. 2018. Recognition from web data: A progressive filtering approach., Vol. 27, 11 (2018), 5303--5315.
[50]
Hantao Yao, Shiliang Zhang, Yongdong Zhang, Jintao Li, and Qi Tian. 2016. Coarse-to-fine description for fine-grained visual categorization., Vol. 25, 10 (2016), 4858--4872.
[51]
Yazhou Yao, Fumin Shen, Guosen Xie, Li Liu, Fan Zhu, Jian Zhang, and Heng Tao Shen. 2020. Exploiting web images for multi-output classification: From category to subcategories. IEEE Transactions on Neural Networks and Learning Systems, Vol. 31, 7 (2020), 2348--2360.
[52]
Yazhou Yao, Fumin Shen, Jian Zhang, Li Liu, Zhenmin Tang, and Ling Shao. 2018a. Extracting multiple visual senses for web learning. IEEE Transactions on Multimedia, Vol. 21, 1 (2018), 184--196.
[53]
Yazhou Yao, Fumin Shen, Jian Zhang, Li Liu, Zhenmin Tang, and Ling Shao. 2018b. Extracting privileged information for enhancing classifier learning. IEEE Transactions on Image Processing, Vol. 28, 1 (2018), 436--450.
[54]
Yazhou Yao, Jian Zhang, Fumin Shen, Xiansheng Hua, Jingsong Xu, and Zhenmin Tang. 2017. Exploiting web images for dataset construction: A domain robust approach. IEEE Transactions on Multimedia, Vol. 19, 8 (2017), 1771--1784.
[55]
Yazhou Yao, Jian Zhang, Fumin Shen, Li Liu, Fan Zhu, Dongxiang Zhang, and Heng Tao Shen. 2019. Towards automatic construction of diverse, high-quality image datasets. IEEE Transactions on Knowledge and Data Engineering, Vol. 32, 6 (2019), 1199--1211.
[56]
Kun Yi and Jianxin Wu. 2019. Probabilistic end-to-end noise correction for learning with noisy labels. In IEEE Conference on Computer Vision and Pattern Recognition. 7017--7025.
[57]
Chiyuan Zhang, Samy Bengio, Moritz Hardt, Benjamin Recht, and Oriol Vinyals. 2017a. Understanding deep learning requires rethinking generalization. In International Conference on Learning Representations.
[58]
Chuanyi Zhang, Yazhou Yao, Huafeng Liu, Guo-Sen Xie, Xiangbo Shu, Tianfei Zhou, Zheng Zhang, Fumin Shen, and Zhenmin Tang. 2020 a. Web-Supervised Network with Softly Update-Drop Training for Fine-Grained Visual Classification. In AAAI Conference on Artificial Intelligence. 12781--12788.
[59]
Chuanyi Zhang, Yazhou Yao, Jiachao Zhang, Jiaxin Chen, and et al. 2020 b. Web-Supervised Network for Fine-Grained Visual Classification. In IEEE International Conference on Multimedia and Expo. 1--6.
[60]
Jian Zhang, Fumin Shen, Xiansheng Hua, Jingsong Xu, and Zhenmin Tang. 2016. Automatic image dataset construction with multiple textual metadata. In IEEE International Conference on Multimedia and Expo. 1--6.
[61]
Jian Zhang, Fumin Shen, Xiansheng Hua, Jingsong Xu, and Zhenmin Tang. 2017b. A new web-supervised method for image dataset constructions. Neurocomputing, Vol. 236 (2017), 23--31.
[62]
Jian Zhang, Fumin Shen, Wankou Yang, Pu Huang, and Zhenmin Tang. 2018. Discovering and distinguishing multiple visual senses for polysemous words. In AAAI Conference on Artificial Intelligence. 523--530.
[63]
Heliang Zheng, Jianlong Fu, Zheng-Jun Zha, and Jiebo Luo. 2019. Looking for the devil in the details: Learning trilinear attention sampling network for fine-grained image recognition. In IEEE Conference on Computer Vision and Pattern Recognition. 5012--5021.
[64]
Bolei Zhou, Aditya Khosla, Agata Lapedriza, Antonio Torralba, and Aude Oliva. 2016. Places: An image database for deep scene understanding. arXiv:1610.02055 (2016).

Cited By

View all
  • (2024)Enhancing Robustness in Learning with Noisy Labels: An Asymmetric Co-Training ApproachProceedings of the 32nd ACM International Conference on Multimedia10.1145/3664647.3680835(4406-4415)Online publication date: 28-Oct-2024
  • (2024)A Survey of Dataset Refinement for Problems in Computer Vision DatasetsACM Computing Surveys10.1145/362715756:7(1-34)Online publication date: 9-Apr-2024
  • (2024)Learning With Imbalanced Noisy Data by Preventing Bias in Sample SelectionIEEE Transactions on Multimedia10.1109/TMM.2024.336891026(7426-7437)Online publication date: 2024
  • Show More Cited By

Index Terms

  1. CRSSC: Salvage Reusable Samples from Noisy Data for Robust Learning

    Recommendations

    Comments

    Please enable JavaScript to view thecomments powered by Disqus.

    Information & Contributors

    Information

    Published In

    cover image ACM Conferences
    MM '20: Proceedings of the 28th ACM International Conference on Multimedia
    October 2020
    4889 pages
    ISBN:9781450379885
    DOI:10.1145/3394171
    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

    Sponsors

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 12 October 2020

    Permissions

    Request permissions for this article.

    Check for updates

    Author Tags

    1. label noise
    2. noisy data
    3. robust learning
    4. sample selection

    Qualifiers

    • Research-article

    Funding Sources

    • Fundamental Research Funds for the Central Universities
    • National Natural Science Foundation of China

    Conference

    MM '20
    Sponsor:

    Acceptance Rates

    Overall Acceptance Rate 2,145 of 8,556 submissions, 25%

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)22
    • Downloads (Last 6 weeks)5
    Reflects downloads up to 10 Nov 2024

    Other Metrics

    Citations

    Cited By

    View all
    • (2024)Enhancing Robustness in Learning with Noisy Labels: An Asymmetric Co-Training ApproachProceedings of the 32nd ACM International Conference on Multimedia10.1145/3664647.3680835(4406-4415)Online publication date: 28-Oct-2024
    • (2024)A Survey of Dataset Refinement for Problems in Computer Vision DatasetsACM Computing Surveys10.1145/362715756:7(1-34)Online publication date: 9-Apr-2024
    • (2024)Learning With Imbalanced Noisy Data by Preventing Bias in Sample SelectionIEEE Transactions on Multimedia10.1109/TMM.2024.336891026(7426-7437)Online publication date: 2024
    • (2024)Learning Discriminative Dynamics with Label Corruption for Noisy Label Detection2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)10.1109/CVPR52733.2024.02121(22477-22487)Online publication date: 16-Jun-2024
    • (2024)Group benefits instance for data purificationComputers and Electrical Engineering10.1016/j.compeleceng.2024.109156116(109156)Online publication date: May-2024
    • (2024)Visual Out-of-Distribution Detection in Open-Set Noisy EnvironmentsInternational Journal of Computer Vision10.1007/s11263-024-02139-y132:11(5453-5470)Online publication date: 16-Jun-2024
    • (2024)HSALC: hard sample aware label correction for medical image classificationMultimedia Tools and Applications10.1007/s11042-024-20114-0Online publication date: 2-Sep-2024
    • (2024)Data reweighting net for web fine-grained image classificationMultimedia Tools and Applications10.1007/s11042-024-18598-x83:33(79985-80005)Online publication date: 2-Mar-2024
    • (2024)Delving Deeper Into Clean Samples for Combating Noisy LabelsPattern Recognition and Computer Vision10.1007/978-981-97-8692-3_13(176-190)Online publication date: 1-Nov-2024
    • (2024)Foster Adaptivity and Balance in Learning with Noisy LabelsComputer Vision – ECCV 202410.1007/978-3-031-73383-3_13(217-235)Online publication date: 3-Nov-2024
    • Show More Cited By

    View Options

    Get Access

    Login options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Media

    Figures

    Other

    Tables

    Share

    Share

    Share this Publication link

    Share on social media