Nothing Special   »   [go: up one dir, main page]

skip to main content
research-article

A hybrid approach with optimization-based and metric-based meta-learner for few-shot learning

Published: 15 July 2019 Publication History

Highlights

We design the meta-metric-learning method that is able to learn task-specific metrics via training a meta-learner.
We propose training methods towards the Meta-Metric-Learner in both single-source and multi-source settings.
We evaluate our model on several benchmark datasets with various baselines and comparative results are given.

Abstract

Few-shot learning aims to learn classifiers for new classes with only a few training examples per class. Most existing few-shot learning approaches belong to either metric-based meta-learning or optimization-based meta-learning category, both of which have achieved successes in the simplified “k-shot N-way” image classification settings. Specifically, the optimization-based approaches train a meta-learner to predict the parameters of the task-specific classifiers. The task-specific classifiers are required to be homogeneous-structured to ease the parameter prediction, so the meta-learning approaches could only handle few-shot learning problems where the tasks share a uniform number of classes. The metric-based approaches learn one task-invariant metric for all the tasks. Even though the metric-learning approaches allow different numbers of classes, they require the tasks all coming from a similar domain such that there exists a uniform metric that could work across tasks. In this work, we propose a hybrid meta-learning model called Meta-Metric-Learner which combines the merits of both optimization- and metric-based approaches. Our meta-metric-learning approach consists of two components, a task-specific metric-based learner as a base model, and a meta-learner that learns and specifies the base model. Thus our model is able to handle flexible numbers of classes as well as generate more generalized metrics for classification across tasks. We test our approach in the standard “k-shot N-way” few-shot learning setting following previous works and a new realistic few-shot setting with flexible class numbers in both single-source form and multi-source form. Experiments show that our approach attains superior performance in all settings.

References

[1]
K. He, X. Zhang, S. Ren, J. Sun, Deep residual learning for image recognition, Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2016, pp. 770–778,.
[2]
A. Krizhevsky, I. Sutskever, G.E. Hinton, Imagenet classification with deep convolutional neural networks, Proceedings of the Advances in Neural Information Processing Systems, 2012, pp. 1097–1105,.
[3]
K. Simonyan, A. Zisserman, Very deep convolutional networks for large-scale image recognition, Proceedings of the International Conference on Learning Representations, 2015.
[4]
L. Fei-Fei, R. Fergus, P. Perona, One-shot learning of object categories, IEEE Trans. Pattern Anal. Mach. Intell. 28 (4) (2006) 594–611,.
[5]
G. Koch, R. Zemel, R. Salakhutdinov, Siamese neural networks for one-shot image recognition, Proceedings of the ICML Deep Learning Workshop, 2, 2015.
[6]
O. Vinyals, C. Blundell, T. Lillicrap, D. Wierstra, et al., Matching networks for one shot learning, Proceedings of the Advances in Neural Information Processing Systems, 2016, pp. 3630–3638.
[7]
Ł. Kaiser, O. Nachum, A. Roy, S. Bengio, Learning to remember rare events, Proceedings of the International Conference on Learning Representations, 2017.
[8]
S. Ravi, H. Larochelle, Optimization as a model for few-shot learning, Proceedings of the International Conference on Learning Representations, 1, 2017.
[9]
T. Munkhdalai, H. Yu, Meta networks, Proceedings of the International Conference on Machine Learning, 2017, pp. 2554–2563.
[10]
Y. Cheng, M. Yu, X. Guo, B. Zhou, Few-shot learning with meta metric learners, Proceedings of the NIPS Workshop on Meta-Learning, 2017.
[11]
E.G. Miller, N.E. Matsakis, P.A. Viola, Learning from one example through shared densities on transforms, Proceedings IEEE Conference on Computer Vision and Pattern Recognition, 1, 2000, pp. 464–471,.
[12]
B.M. Lake, R. Salakhutdinov, J.B. Tenenbaum, Human-level concept learning through probabilistic program induction, Science 350 (6266) (2015) 1332–1338,.
[13]
S. Thrun, L. Pratt, Learning to learn: introduction and overview, Learning to Learn, Springer, 1998, pp. 3–17,.
[14]
J. Schmidhuber, J. Zhao, M. Wiering, Shifting inductive bias with success-story algorithm, adaptive levin search, and incremental self-improvement, Mach. Learn. 28 (1) (1997) 105–130,.
[15]
A. Santoro, S. Bartunov, M. Botvinick, D. Wierstra, T. Lillicrap, Meta-Learning with Memory-Augmented Neural Networks, Proceedings of the International Conference on Machine Learning, 2016, pp. 1842–1850.
[16]
A. Graves, G. Wayne, I. Danihelka, Neural turing machines, 2014.
[17]
J. Snell, K. Swersky, R. Zemel, Prototypical networks for few-shot learning, Proceedings of the Advances in Neural Information Processing Systems, 2017, pp. 4077–4087.
[18]
A. Mehrotra, A. Dukkipati, Generative adversarial residual pairwise networks for one shot learning, 2017.
[19]
F. Sung, Y. Yang, L. Zhang, T. Xiang, P.H. Torr, T.M. Hospedales, Learning to compare: relation network for few-shot learning, Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2018, pp. 1199–1208,.
[20]
M. Andrychowicz, M. Denil, S. Gomez, M.W. Hoffman, D. Pfau, T. Schaul, N. de Freitas, Learning to learn by gradient descent by gradient descent, Proceedings of the Advances in Neural Information Processing Systems, 2016, pp. 3981–3989.
[21]
C. Finn, P. Abbeel, S. Levine, Model-agnostic meta-learning for fast adaptation of deep networks, Proceedings of the Thirty Fourth International Conference on Machine Learning, 70, 2017, pp. 1126–1135.
[22]
Z. Li, F. Zhou, F. Chen, H. Li, Meta-SGD: learning to learn quickly for few-shot learning, 2017.
[23]
M. Abdullah Jamal, G.-J. Qi, M. Shah, Task-agnostic meta-learning for few-shot learning, 2018.
[24]
C. Finn, K. Xu, S. Levine, Probabilistic model-agnostic meta-learning, Proceedings of the Advances in Neural Information Processing Systems, 2018, pp. 9537–9548.
[25]
J. Yoon, T. Kim, O. Dia, S. Kim, Y. Bengio, S. Ahn, Bayesian model-agnostic meta-learning, Proceedings of the Advances in Neural Information Processing Systems, 2018, pp. 7343–7353.
[26]
F. Zhou, B. Wu, Z. Li, Deep meta-learning: learning to learn in the concept space, 2018.
[27]
V.G. Satorras, J.B. Estrach, Few-shot learning with graph neural networks, Proceedings of the International Conference on Learning Representations, 2018.
[28]
Z. Chen, Y. Fu, Y. Zhang, L. Sigal, Multi-level semantic feature augmentation for one-shot learning, IEEE Transactions on Image Processing, 2019,.
[29]
B. Oreshkin, P. Rodríguez López, A. Lacoste, Tadam: task dependent adaptive metric for improved few-shot learning, Proceedings of the Advances in Neural Information Processing Systems, 2018, pp. 719–729.
[30]
A.A. Rusu, D. Rao, J. Sygnowski, O. Vinyals, R. Pascanu, S. Osindero, R. Hadsell, Meta-learning with latent embedding optimization, Proceedings of the International Conference on Learning Representations, 2019.
[31]
T. Scott, K. Ridgeway, M.C. Mozer, Adapted deep embeddings: a synthesis of methods for k-shot inductive transfer learning, Proceedings of the Advances in Neural Information Processing Systems, 2018, pp. 76–85.
[32]
M. Yu, X. Guo, J. Yi, S. Chang, S. Potdar, Y. Cheng, G. Tesauro, H. Wang, B. Zhou, Diverse few-shot text classification with multiple metrics, Proceedings of the Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long Papers), 2018, pp. 1206–1215,.
[33]
G. Griffin, A. Holub, P. Perona, Caltech-256 object category dataset, Technical Report 7694, California Institute of Technology, 2007.
[34]
A. Krizhevsky, G. Hinton, Learning multiple layers of features from tiny images, Technical Report, Citeseer, 2009.
[35]
C. Wah, S. Branson, P. Welinder, P. Perona, S. Belongie, The Caltech-UCSD birds-200-2011 dataset, Technical Report CNS-TR-2011-001, California Institute of Technology, 2011.

Cited By

View all

Index Terms

  1. A hybrid approach with optimization-based and metric-based meta-learner for few-shot learning
        Index terms have been assigned to the content through auto-classification.

        Recommendations

        Comments

        Please enable JavaScript to view thecomments powered by Disqus.

        Information & Contributors

        Information

        Published In

        cover image Neurocomputing
        Neurocomputing  Volume 349, Issue C
        Jul 2019
        327 pages

        Publisher

        Elsevier Science Publishers B. V.

        Netherlands

        Publication History

        Published: 15 July 2019

        Author Tags

        1. Few-shot learning
        2. Meta-learning
        3. Image classification
        4. Meta-metric-learner

        Qualifiers

        • Research-article

        Contributors

        Other Metrics

        Bibliometrics & Citations

        Bibliometrics

        Article Metrics

        • Downloads (Last 12 months)0
        • Downloads (Last 6 weeks)0
        Reflects downloads up to 23 Nov 2024

        Other Metrics

        Citations

        Cited By

        View all
        • (2024)MERA: Meta-Learning Based Runtime Adaptation for Industrial Wireless Sensor-Actuator NetworksACM Transactions on Sensor Networks10.1145/366533020:4(1-24)Online publication date: 8-Jul-2024
        • (2024)Advances and Challenges in Meta-Learning: A Technical ReviewIEEE Transactions on Pattern Analysis and Machine Intelligence10.1109/TPAMI.2024.335784746:7(4763-4779)Online publication date: 24-Jan-2024
        • (2024)FTANImage and Vision Computing10.1016/j.imavis.2024.105159149:COnline publication date: 18-Oct-2024
        • (2023)MCI-HyperNetNeurocomputing10.1016/j.neucom.2023.126606554:COnline publication date: 14-Oct-2023
        • (2023)Training with scaled logits to alleviate class-level over-fitting in few-shot learningNeurocomputing10.1016/j.neucom.2022.12.011522:C(142-151)Online publication date: 14-Feb-2023
        • (2023)ReNAPNeurocomputing10.1016/j.neucom.2022.11.082520:C(356-364)Online publication date: 1-Feb-2023
        • (2023)A novel hybrid meta-learning for few-shot gas-insulated switchgear insulation defect diagnosisExpert Systems with Applications: An International Journal10.1016/j.eswa.2023.120956233:COnline publication date: 15-Dec-2023
        • (2023)Multiview meta-metric learning for sign language recognition using triplet loss embeddingsPattern Analysis & Applications10.1007/s10044-023-01134-226:3(1125-1141)Online publication date: 1-Aug-2023
        • (2022)Metrics Space and NormScientific Programming10.1155/2022/19113452022Online publication date: 1-Jan-2022
        • (2022)A New Similarity Space Tailored for Supervised Deep Metric LearningACM Transactions on Intelligent Systems and Technology10.1145/355976614:1(1-25)Online publication date: 9-Nov-2022
        • Show More Cited By

        View Options

        View options

        Login options

        Media

        Figures

        Other

        Tables

        Share

        Share

        Share this Publication link

        Share on social media