Abstract
Few-shot learning is used in the development of models that can acquire novel class concepts from limited training samples, facilitating rapid adaptation to novel, intricate, and varied tasks encountered in real-world scenarios. Compared to meta-training, traditional batch training exhibits superior efficiency. However, when addressing few-shot learning tasks, the features extracted by batch-trained models often struggle to effectively represent novel class concepts, resulting in unsatisfactory performance. This challenge arises from two primary issues. First, the models lack the capacity to autonomously map novel class features into an effective discriminant subspace, resulting in the interference of class-independent feature components during metric calculations. Second, the limited sample size hinders the ability of the model to accumulate valuable experience, leading to representations that deviate from the true class center. To address these issues, we introduce the Class Feature Sub-space (CFS-space) as an effective discriminant space for few-shot classification; it preserves class features and suppresses noise components by mapping the extracted features into the CFS-space. Furthermore, we incorporate empirical knowledge from the base set and calibrate the prototypes within the CFS-Space to enhance the class representations. Ablation studies affirm the efficacy of our approach. As evaluated in 5-way 1/5-shot tasks, our method achieves impressive accuracies of 66.42%/83.69% on mini-ImageNet, 72.07%/86.36% on tiered-ImageNet, and 79.34%/90.42% on CUB, significantly narrowing the performance gap between the batch-training and the meta-training paradigm.
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.Data Availability
All the data used in this paper are publicly available. No ethical approval or informed consent was necessary for this study, as the data were already publicly available and did not involve human or animal subjects.
References
Zheng S, Zhang Y, Liu W, Zou Y (2020) Improved image representation and sparse representation for image classification. Appl Intell 50:1687–1698
Ren J, Shi M, Chen J, Wang R, Wang X (2022) Hyperspectral image classification using multi-level features fusion capsule network with a dense structure. Appl Intell 1–20
Hudson DA, Zitnick L (2021) Generative adversarial transformers. In: International conference on machine learning, pp 4487–4499. PMLR
Zhu X, Su W, Lu L, Li B, Wang X, Dai J (2020) Deformable detr: deformable transformers for end-to-end object detection. In: International conference on learning representations
Pal SK, Pramanik A, Maiti J, Mitra P (2021) Deep learning in multi-object detection and tracking: state of the art. Appl Intell 51:6400–6429
Zhang J, Liu Y, Guo C, Zhan J (2023) Optimized segmentation with image inpainting for semantic mapping in dynamic scenes. Appl Intell 53(2):2173–2188
Hou C, Zhang W, Wang H, Liu F, Liu D, Chang J (2022) A semantic segmentation model for lumbar mri images using divergence loss. Appl Intell 1–14
Wang W, Xia Q, Hu Z, Yan Z, Li Z, Wu Y, Huang N, Gao Y, Metaxas D, Zhang S (2021) Few-shot learning by a cascaded framework with shape-constrained pseudo label assessment for whole heart segmentation. IEEE Trans Med Imaging 40(10):2629–2641
Ouyang C, Biffi C, Chen C, Kart T, Qiu H, Rueckert D (2020) Self-supervision with superpixels: training few-shot medical image segmentation without annotation. In: European conference on computer vision, pp 762–780. Springer
Kuchuk H, Podorozhniak A, Hlavcheva D, Yaloveha V (2020) Application of deep learning in the processing of the aerospace system’s multispectral images. In: Handbook of research on artificial intelligence applications in the aviation and aerospace industries, pp 134–147. IGI Global
Zeng W, Quan Z, Zhao Z, Xie C, Lu X (2020) A deep learning approach for aircraft trajectory prediction in terminal airspace. IEEE Access 8:151250–151266
Wu J, Zhao Z, Sun C, Yan R, Chen X (2020) Few-shot transfer learning for intelligent fault diagnosis of machine. Measurement 166:108202
Zhou X, Liang W, Shimizu S, Ma J, Jin Q (2020) Siamese neural network based few-shot learning for anomaly detection in industrial cyber-physical systems. IEEE Trans Ind Inform 17(8):5790–5798
Wang Y, Yao Q, Kwok JT, Ni LM (2020) Generalizing from a few examples: a survey on few-shot learning. ACM Comput Surv (CSUR) 53(3):1–34
Snell J, Swersky K, Zemel R (2017) Prototypical networks for few-shot learning. Adv Neural Inf Process Syst 30:1
Zhang C, Cai Y, Lin G, Shen C (2022) Deepemd: differentiable earth mover’s distance for few-shot learning. IEEE Trans Pattern Anal Mach Intell 45(5):5632–5648
Wertheimer D, Tang L, Hariharan B (2021) Few-shot classification with feature map reconstruction networks. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 8012–8021
Xie J, Long F, Lv J, Wang Q, Li P (2022) Joint distribution matters: deep brownian distance covariance for few-shot classification. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 7972–7981
Yang S, Liu L, Xu M (2020) Free lunch for few-shot learning: distribution calibration. In: International conference on learning representations
Tian Y, Wang Y, Krishnan D, Tenenbaum JB, Isola P (2020) Rethinking few-shot image classification: a good embedding is all you need? In: European conference on computer vision, pp 266–282. Springer
Mangla P, Kumari N, Sinha A, Singh M, Krishnamurthy B, Balasubramanian VN (2020) Charting the right manifold: manifold mixup for few-shot learning. In: Proceedings of the IEEE/CVF winter conference on applications of computer vision, pp 2218–2227
Vinyals O, Blundell C, Lillicrap T, Wierstra D et al (2016) Matching networks for one shot learning. Adv Neural Inf Process Syst 29
Finn C, Abbeel P, Levine S (2017) Model-agnostic meta-learning for fast adaptation of deep networks. In: International conference on machine learning, pp 1126–1135. PMLR
Simon C, Koniusz P, Nock R, Harandi M (2020) Adaptive subspaces for few-shot learning. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 4136–4145
Devos A, Grossglauser M (2020) Regression networks for meta-learning few-shot classification. In: 7th ICML Workshop on automated machine learning (AutoML 2020)
Shao Y, Wu W, You X, Gao C, Sang N (2022) Improving the generalization of maml in few-shot classification via bi-level constraint. IEEE Trans Circ Syst Vid Technol
Zhu X, Li S (2022) Mgml: momentum group meta-learning for few-shot image classification. Neurocomputing 514:351–361
Fang C, He H, Long Q, Su WJ (2021) Exploring deep neural networks via layer-peeled model: minority collapse in imbalanced training. Proc Natl Acad Sci 118(43):2103091118
Papyan V, Han X, Donoho DL (2020) Prevalence of neural collapse during the terminal phase of deep learning training. Proc Natl Acad Sci 117(40):24652–24663
Hou R, Chang H, Ma B, Shan S, Chen X (2019) Cross attention network for few-shot classification. Adv Neural Inf Process Syst 32
Li W, Wang Z, Yang X, Dong C, Tian P, Qin T, Huo J, Shi Y, Wang L, Gao Y et al (2023) Libfewshot: a comprehensive library for few-shot learning. IEEE Trans Pattern Anal Mach Intell
Afrasiyabi A, Larochelle H, Lalonde J-F, Gagné C (2022) Matching feature sets for few-shot image classification. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 9014–9024
Yoon SW, Seo J, Moon J (2019) Tapnet: neural network augmented with task-adaptive projection for few-shot learning. In: International conference on machine learning, pp 7115–7123. PMLR
Li L, Ge H, Gao J, Zhang Y, Tong Y, Sun J (2020) A novel geometric mean feature space discriminant analysis method for hyperspectral image feature extraction. Neural Process Lett 51(1):515–542
Xing C, Rostamzadeh N, Oreshkin B, O Pinheiro PO (2019) Adaptive cross-modal few-shot learning. Adv Neural Inf Process Syst 32:1
Xu J, Le H (2022) Generating representative samples for few-shot classification. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 9003–9013
Zhang B, Li X, Ye Y, Huang Z, Zhang L (2021) Prototype completion with primitive knowledge for few-shot learning. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 3754–3762
Radford A, Kim JW, Hallacy C, Ramesh A, Goh G, Agarwal S, Sastry G, Askell A, Mishkin P, Clark J et al (2021) Learning transferable visual models from natural language supervision. In: International conference on machine learning, pp 8748–8763. PMLR
Chen W-Y, Liu Y-C, Kira Z, Wang Y-CF, Huang J-B (2019) A closer look at few-shot classification. In: International conference on learning representations
Jiang W, Huang K, Geng J, Deng X (2020) Multi-scale metric learning for few-shot learning. IEEE Trans Circ Syst Vid Technol 31(3):1091–1102
Allen K, Shelhamer E, Shin H, Tenenbaum J (2019) Infinite mixture prototypes for few-shot learning. In: International conference on machine learning, pp 232–241. PMLR
Ren M, Triantafillou E, Ravi S, Snell J, Swersky K, Tenenbaum JB, Larochelle H, Zemel RS (2018) Meta-learning for semi-supervised few-shot classification. In: Proceedings of 6th International Conference on Learning Representations ICLR
Wah C, Branson S, Welinder P, Perona P, Belongie S (2011) The caltech-ucsd birds-200-2011 dataset
Snell J, Zemel R (2020) Bayesian few-shot classification with one-vs-each pólya-gamma augmented gaussian processes. In: International conference on learning representations
Oh J, Yoo H, Kim C, Yun S (2021) Boil: Towards representation change for few-shot learning. In: The Ninth international conference on learning representations (ICLR). The International Conference on Learning Representations (ICLR)
Oh J, Yoo H, Kim C, Yun S (2021) Boil: towards representation change for few-shot learning. In: The Ninth International Conference on Learning Representations (ICLR). The International Conference on Learning Representations (ICLR)
Lee K, Maji S, Ravichandran A, Soatto S (2019) Meta-learning with differentiable convex optimization. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 10657–10665
Li H, Eigen D, Dodge S, Zeiler M, Wang X (2019) Finding task-relevant features for few-shot learning by category traversal. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 1–10
Chen Y, Liu Z, Xu H, Darrell T, Wang X (2021) Meta-baseline: exploring simple meta-learning for few-shot learning. In: Proceedings of the IEEE/CVF international conference on computer vision, pp 9062–9071
Chen Z, Ge J, Zhan H, Huang S, Wang D (2021) Pareto self-supervised training for few-shot learning. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 13663–13672
Liu Y, Zhang W, Xiang C, Zheng T, Cai D, He X (2022) Learning to affiliate: mutual centralized learning for few-shot classification. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 14411–14420
Hao F, He F, Cheng J, Tao D (2021) Global-local interplay in semantic alignment for few-shot learning. IEEE Trans Circ Syst Vid Technol 32(7):4351–4363
Zhang J, Zhang X, Wang Z (2022) Task encoding with distribution calibration for few-shot learning. IEEE Trans Circ Syst Vid Technol 32(9):6240–6252
He X, Lin J (2022) Weakly-supervised object localization based fine-grained few-shot learning. J Image Graph (007):027
Zhang M, Huang S, Li W, Wang D (2022) Tree structure-aware few-shot image classification via hierarchical aggregation. In: European conference on computer vision, pp 453–470. Springer
Wang L, He K, Liu Z (2024) Mcs: a metric confidence selection framework for few shot image classification. Multimed Tools Appl 83(4):10865–10880
Liu B, Cao Y, Lin Y, Li Q, Zhang Z, Long M, Hu H (2020) Negative margin matters: understanding margin in few-shot classification. In: European conference on computer vision, pp 438–455. Springer
Cheng J, Hao F, Liu L, Tao D (2022) Imposing semantic consistency of local descriptors for few-shot learning. IEEE Trans Image Process 31:1587–1600
Goldblum M, Reich S, Fowl L, Ni R, Cherepanova V, Goldstein T (2020) Unraveling meta-learning: understanding feature representations for few-shot tasks. In: International conference on machine learning, pp 3607–3616. PMLR
Hossain MM, Walid MAA, Galib SS, Azad MM, Rahman W, Shafi A, Rahman MM (2024) Covid-19 detection from chest ct images using optimized deep features and ensemble classification. Syst Soft Comput, 200077
Biagetti G, Crippa P, Falaschetti L, Luzzi S, Turchetti C (2021) Classification of alzheimer’s disease from eeg signal using robust-pca feature extraction. Procedia Comput Sci 192:3114–3122
Author information
Authors and Affiliations
Contributions
Bin Song was involved in the design, implementation, formal analysis and writing. Hong Zhu provided guidance, project administration, and supervision. Bingxin Wang conducted review and editing. Yuandong Bi was responsible for validation.
Corresponding author
Ethics declarations
Competing Interest
The paper is original in its contents and is not under consideration for publication in any other journals/proceedings. The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Song, B., Zhu, H., Wang, B. et al. Class feature Sub-space for few-shot classification. Appl Intell 54, 9177–9194 (2024). https://doi.org/10.1007/s10489-024-05635-3
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10489-024-05635-3