Nothing Special   »   [go: up one dir, main page]

Skip to main content
Log in

Class feature Sub-space for few-shot classification

  • Published:
Applied Intelligence Aims and scope Submit manuscript

Abstract

Few-shot learning is used in the development of models that can acquire novel class concepts from limited training samples, facilitating rapid adaptation to novel, intricate, and varied tasks encountered in real-world scenarios. Compared to meta-training, traditional batch training exhibits superior efficiency. However, when addressing few-shot learning tasks, the features extracted by batch-trained models often struggle to effectively represent novel class concepts, resulting in unsatisfactory performance. This challenge arises from two primary issues. First, the models lack the capacity to autonomously map novel class features into an effective discriminant subspace, resulting in the interference of class-independent feature components during metric calculations. Second, the limited sample size hinders the ability of the model to accumulate valuable experience, leading to representations that deviate from the true class center. To address these issues, we introduce the Class Feature Sub-space (CFS-space) as an effective discriminant space for few-shot classification; it preserves class features and suppresses noise components by mapping the extracted features into the CFS-space. Furthermore, we incorporate empirical knowledge from the base set and calibrate the prototypes within the CFS-Space to enhance the class representations. Ablation studies affirm the efficacy of our approach. As evaluated in 5-way 1/5-shot tasks, our method achieves impressive accuracies of 66.42%/83.69% on mini-ImageNet, 72.07%/86.36% on tiered-ImageNet, and 79.34%/90.42% on CUB, significantly narrowing the performance gap between the batch-training and the meta-training paradigm.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Algorithm 1
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8

Similar content being viewed by others

Explore related subjects

Discover the latest articles, news and stories from top researchers in related subjects.

Data Availability

All the data used in this paper are publicly available. No ethical approval or informed consent was necessary for this study, as the data were already publicly available and did not involve human or animal subjects.

References

  1. Zheng S, Zhang Y, Liu W, Zou Y (2020) Improved image representation and sparse representation for image classification. Appl Intell 50:1687–1698

    Article  Google Scholar 

  2. Ren J, Shi M, Chen J, Wang R, Wang X (2022) Hyperspectral image classification using multi-level features fusion capsule network with a dense structure. Appl Intell 1–20

  3. Hudson DA, Zitnick L (2021) Generative adversarial transformers. In: International conference on machine learning, pp 4487–4499. PMLR

  4. Zhu X, Su W, Lu L, Li B, Wang X, Dai J (2020) Deformable detr: deformable transformers for end-to-end object detection. In: International conference on learning representations

  5. Pal SK, Pramanik A, Maiti J, Mitra P (2021) Deep learning in multi-object detection and tracking: state of the art. Appl Intell 51:6400–6429

    Article  Google Scholar 

  6. Zhang J, Liu Y, Guo C, Zhan J (2023) Optimized segmentation with image inpainting for semantic mapping in dynamic scenes. Appl Intell 53(2):2173–2188

    Article  Google Scholar 

  7. Hou C, Zhang W, Wang H, Liu F, Liu D, Chang J (2022) A semantic segmentation model for lumbar mri images using divergence loss. Appl Intell 1–14

  8. Wang W, Xia Q, Hu Z, Yan Z, Li Z, Wu Y, Huang N, Gao Y, Metaxas D, Zhang S (2021) Few-shot learning by a cascaded framework with shape-constrained pseudo label assessment for whole heart segmentation. IEEE Trans Med Imaging 40(10):2629–2641

    Article  Google Scholar 

  9. Ouyang C, Biffi C, Chen C, Kart T, Qiu H, Rueckert D (2020) Self-supervision with superpixels: training few-shot medical image segmentation without annotation. In: European conference on computer vision, pp 762–780. Springer

  10. Kuchuk H, Podorozhniak A, Hlavcheva D, Yaloveha V (2020) Application of deep learning in the processing of the aerospace system’s multispectral images. In: Handbook of research on artificial intelligence applications in the aviation and aerospace industries, pp 134–147. IGI Global

  11. Zeng W, Quan Z, Zhao Z, Xie C, Lu X (2020) A deep learning approach for aircraft trajectory prediction in terminal airspace. IEEE Access 8:151250–151266

    Article  Google Scholar 

  12. Wu J, Zhao Z, Sun C, Yan R, Chen X (2020) Few-shot transfer learning for intelligent fault diagnosis of machine. Measurement 166:108202

    Article  Google Scholar 

  13. Zhou X, Liang W, Shimizu S, Ma J, Jin Q (2020) Siamese neural network based few-shot learning for anomaly detection in industrial cyber-physical systems. IEEE Trans Ind Inform 17(8):5790–5798

    Article  Google Scholar 

  14. Wang Y, Yao Q, Kwok JT, Ni LM (2020) Generalizing from a few examples: a survey on few-shot learning. ACM Comput Surv (CSUR) 53(3):1–34

  15. Snell J, Swersky K, Zemel R (2017) Prototypical networks for few-shot learning. Adv Neural Inf Process Syst 30:1

    Google Scholar 

  16. Zhang C, Cai Y, Lin G, Shen C (2022) Deepemd: differentiable earth mover’s distance for few-shot learning. IEEE Trans Pattern Anal Mach Intell 45(5):5632–5648

  17. Wertheimer D, Tang L, Hariharan B (2021) Few-shot classification with feature map reconstruction networks. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 8012–8021

  18. Xie J, Long F, Lv J, Wang Q, Li P (2022) Joint distribution matters: deep brownian distance covariance for few-shot classification. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 7972–7981

  19. Yang S, Liu L, Xu M (2020) Free lunch for few-shot learning: distribution calibration. In: International conference on learning representations

  20. Tian Y, Wang Y, Krishnan D, Tenenbaum JB, Isola P (2020) Rethinking few-shot image classification: a good embedding is all you need? In: European conference on computer vision, pp 266–282. Springer

  21. Mangla P, Kumari N, Sinha A, Singh M, Krishnamurthy B, Balasubramanian VN (2020) Charting the right manifold: manifold mixup for few-shot learning. In: Proceedings of the IEEE/CVF winter conference on applications of computer vision, pp 2218–2227

  22. Vinyals O, Blundell C, Lillicrap T, Wierstra D et al (2016) Matching networks for one shot learning. Adv Neural Inf Process Syst 29

  23. Finn C, Abbeel P, Levine S (2017) Model-agnostic meta-learning for fast adaptation of deep networks. In: International conference on machine learning, pp 1126–1135. PMLR

  24. Simon C, Koniusz P, Nock R, Harandi M (2020) Adaptive subspaces for few-shot learning. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 4136–4145

  25. Devos A, Grossglauser M (2020) Regression networks for meta-learning few-shot classification. In: 7th ICML Workshop on automated machine learning (AutoML 2020)

  26. Shao Y, Wu W, You X, Gao C, Sang N (2022) Improving the generalization of maml in few-shot classification via bi-level constraint. IEEE Trans Circ Syst Vid Technol

  27. Zhu X, Li S (2022) Mgml: momentum group meta-learning for few-shot image classification. Neurocomputing 514:351–361

  28. Fang C, He H, Long Q, Su WJ (2021) Exploring deep neural networks via layer-peeled model: minority collapse in imbalanced training. Proc Natl Acad Sci 118(43):2103091118

  29. Papyan V, Han X, Donoho DL (2020) Prevalence of neural collapse during the terminal phase of deep learning training. Proc Natl Acad Sci 117(40):24652–24663

    Article  MathSciNet  Google Scholar 

  30. Hou R, Chang H, Ma B, Shan S, Chen X (2019) Cross attention network for few-shot classification. Adv Neural Inf Process Syst 32

  31. Li W, Wang Z, Yang X, Dong C, Tian P, Qin T, Huo J, Shi Y, Wang L, Gao Y et al (2023) Libfewshot: a comprehensive library for few-shot learning. IEEE Trans Pattern Anal Mach Intell

  32. Afrasiyabi A, Larochelle H, Lalonde J-F, Gagné C (2022) Matching feature sets for few-shot image classification. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 9014–9024

  33. Yoon SW, Seo J, Moon J (2019) Tapnet: neural network augmented with task-adaptive projection for few-shot learning. In: International conference on machine learning, pp 7115–7123. PMLR

  34. Li L, Ge H, Gao J, Zhang Y, Tong Y, Sun J (2020) A novel geometric mean feature space discriminant analysis method for hyperspectral image feature extraction. Neural Process Lett 51(1):515–542

    Article  Google Scholar 

  35. Xing C, Rostamzadeh N, Oreshkin B, O Pinheiro PO (2019) Adaptive cross-modal few-shot learning. Adv Neural Inf Process Syst 32:1

    Google Scholar 

  36. Xu J, Le H (2022) Generating representative samples for few-shot classification. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 9003–9013

  37. Zhang B, Li X, Ye Y, Huang Z, Zhang L (2021) Prototype completion with primitive knowledge for few-shot learning. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 3754–3762

  38. Radford A, Kim JW, Hallacy C, Ramesh A, Goh G, Agarwal S, Sastry G, Askell A, Mishkin P, Clark J et al (2021) Learning transferable visual models from natural language supervision. In: International conference on machine learning, pp 8748–8763. PMLR

  39. Chen W-Y, Liu Y-C, Kira Z, Wang Y-CF, Huang J-B (2019) A closer look at few-shot classification. In: International conference on learning representations

  40. Jiang W, Huang K, Geng J, Deng X (2020) Multi-scale metric learning for few-shot learning. IEEE Trans Circ Syst Vid Technol 31(3):1091–1102

    Article  Google Scholar 

  41. Allen K, Shelhamer E, Shin H, Tenenbaum J (2019) Infinite mixture prototypes for few-shot learning. In: International conference on machine learning, pp 232–241. PMLR

  42. Ren M, Triantafillou E, Ravi S, Snell J, Swersky K, Tenenbaum JB, Larochelle H, Zemel RS (2018) Meta-learning for semi-supervised few-shot classification. In: Proceedings of 6th International Conference on Learning Representations ICLR

  43. Wah C, Branson S, Welinder P, Perona P, Belongie S (2011) The caltech-ucsd birds-200-2011 dataset

  44. Snell J, Zemel R (2020) Bayesian few-shot classification with one-vs-each pólya-gamma augmented gaussian processes. In: International conference on learning representations

  45. Oh J, Yoo H, Kim C, Yun S (2021) Boil: Towards representation change for few-shot learning. In: The Ninth international conference on learning representations (ICLR). The International Conference on Learning Representations (ICLR)

  46. Oh J, Yoo H, Kim C, Yun S (2021) Boil: towards representation change for few-shot learning. In: The Ninth International Conference on Learning Representations (ICLR). The International Conference on Learning Representations (ICLR)

  47. Lee K, Maji S, Ravichandran A, Soatto S (2019) Meta-learning with differentiable convex optimization. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 10657–10665

  48. Li H, Eigen D, Dodge S, Zeiler M, Wang X (2019) Finding task-relevant features for few-shot learning by category traversal. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 1–10

  49. Chen Y, Liu Z, Xu H, Darrell T, Wang X (2021) Meta-baseline: exploring simple meta-learning for few-shot learning. In: Proceedings of the IEEE/CVF international conference on computer vision, pp 9062–9071

  50. Chen Z, Ge J, Zhan H, Huang S, Wang D (2021) Pareto self-supervised training for few-shot learning. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 13663–13672

  51. Liu Y, Zhang W, Xiang C, Zheng T, Cai D, He X (2022) Learning to affiliate: mutual centralized learning for few-shot classification. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 14411–14420

  52. Hao F, He F, Cheng J, Tao D (2021) Global-local interplay in semantic alignment for few-shot learning. IEEE Trans Circ Syst Vid Technol 32(7):4351–4363

    Article  Google Scholar 

  53. Zhang J, Zhang X, Wang Z (2022) Task encoding with distribution calibration for few-shot learning. IEEE Trans Circ Syst Vid Technol 32(9):6240–6252

    Article  Google Scholar 

  54. He X, Lin J (2022) Weakly-supervised object localization based fine-grained few-shot learning. J Image Graph (007):027

  55. Zhang M, Huang S, Li W, Wang D (2022) Tree structure-aware few-shot image classification via hierarchical aggregation. In: European conference on computer vision, pp 453–470. Springer

  56. Wang L, He K, Liu Z (2024) Mcs: a metric confidence selection framework for few shot image classification. Multimed Tools Appl 83(4):10865–10880

    Article  Google Scholar 

  57. Liu B, Cao Y, Lin Y, Li Q, Zhang Z, Long M, Hu H (2020) Negative margin matters: understanding margin in few-shot classification. In: European conference on computer vision, pp 438–455. Springer

  58. Cheng J, Hao F, Liu L, Tao D (2022) Imposing semantic consistency of local descriptors for few-shot learning. IEEE Trans Image Process 31:1587–1600

    Article  Google Scholar 

  59. Goldblum M, Reich S, Fowl L, Ni R, Cherepanova V, Goldstein T (2020) Unraveling meta-learning: understanding feature representations for few-shot tasks. In: International conference on machine learning, pp 3607–3616. PMLR

  60. Hossain MM, Walid MAA, Galib SS, Azad MM, Rahman W, Shafi A, Rahman MM (2024) Covid-19 detection from chest ct images using optimized deep features and ensemble classification. Syst Soft Comput, 200077

  61. Biagetti G, Crippa P, Falaschetti L, Luzzi S, Turchetti C (2021) Classification of alzheimer’s disease from eeg signal using robust-pca feature extraction. Procedia Comput Sci 192:3114–3122

Download references

Author information

Authors and Affiliations

Authors

Contributions

Bin Song was involved in the design, implementation, formal analysis and writing. Hong Zhu provided guidance, project administration, and supervision. Bingxin Wang conducted review and editing. Yuandong Bi was responsible for validation.

Corresponding author

Correspondence to Hong Zhu.

Ethics declarations

Competing Interest

The paper is original in its contents and is not under consideration for publication in any other journals/proceedings. The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Song, B., Zhu, H., Wang, B. et al. Class feature Sub-space for few-shot classification. Appl Intell 54, 9177–9194 (2024). https://doi.org/10.1007/s10489-024-05635-3

Download citation

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s10489-024-05635-3

Keywords

Navigation