Nothing Special   »   [go: up one dir, main page]

skip to main content
10.1145/3394486.3406466acmconferencesArticle/Chapter ViewAbstractPublication PageskddConference Proceedingsconference-collections
tutorial

Learning with Small Data

Published: 20 August 2020 Publication History

Abstract

In the era of big data, data-driven methods have become increasingly popular in various applications, such as image recognition, traffic signal control, fake news detection. The superior performance of these data-driven approaches relies on large-scale labeled training data, which are probably inaccessible in real-world applications, i.e., "small (labeled) data" challenge. Examples include predicting emergent events in a city, detecting emerging fake news, and forecasting the progression of conditions for rare diseases. In most scenarios, people care about these small data cases most and thus improving the learning effectiveness of machine learning algorithms with small labeled data has been a popular research topic.
In this tutorial, we will review the trending state-of-the-art machine learning techniques for learning with small (labeled) data. These techniques are organized from two aspects: (1) providing a comprehensive review of recent studies about knowledge generalization, transfer, and sharing, where transfer learning, multi-task learning, and meta-learning are discussed. Particularly, we will focus more on meta-learning, which improves the model generalization ability and has been proven to be an effective approach recently; (2) introducing the cutting-edge techniques which focus on incorporating domain knowledge into machine learning models. Different from model-based knowledge transfer techniques, in real-world applications, domain knowledge (e.g., physical laws) provides us with a new angle to deal with the small data challenge. Specifically, domain knowledge can be used to optimize learning strategies and/or guide the model design. In data mining field, we believe that learning with small data is a trending topic with important social impact, which will attract both researchers and practitioners from academia and industry.

References

[1]
Anurag Ajay, Jiajun Wu, Nima Fazeli, Maria Bauza, Leslie P Kaelbling, Joshua B Tenenbaum, and Alberto Rodriguez. 2018. Augmenting physical simulators with stochastic neural networks: Case study of planar pushing and bouncing. In IROS.
[2]
Y Ba, G Zhao, and A Kadambi. 2019. Blending diverse physical priors with neural networks. arXiv:1910.00201 (2019).
[3]
Hakan Bilen and Andrea Vedaldi. 2016. Integrated perception with recurrent multi-task neural networks. In NeurIPS. 235--243.
[4]
Avishek Joey Bose, Ankit Jain, Piero Molino, and William L Hamilton. 2019. Meta-Graph: Few shot Link Prediction via Meta Learning. arXiv preprint arXiv:1912.09867 (2019).
[5]
Chelsea Finn, Pieter Abbeel, and Sergey Levine. 2017. Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks. In ICML. 1126--1135.
[6]
Mingsheng Long, Yue Cao, Jianmin Wang, and Michael I Jordan. 2015. Learning transferable features with deep adaptation networks. In ICML.
[7]
Ishan Misra, Abhinav Shrivastava, Abhinav Gupta, and Martial Hebert. 2016. Cross-stitch networks for multi-task learning. In CVPR. 3994--4003.
[8]
Tsendsuren Munkhdalai, Xingdi Yuan, Soroush Mehri, and Adam Trischler. 2018. Rapid adaptation with conditionally shifted neurons. In ICML. 3661--3670.
[9]
Nikhil Muralidhar, Jie Bu, Ze Cao, Long He, Naren Ramakrishnan, Danesh Tafti, and Anuj Karpatne. 2019. Physics-guided Design and Learning of Neural Networks for Predicting Drag Force on Particle Suspensions in Moving Fluids. arXiv preprint arXiv:1911.04240 (2019).
[10]
L Ruthotto and E Haber. 2018. Deep neural networks motivated by partial differential equations. Journal of Mathematical Imaging and Vision (2018).
[11]
Jake Snell, Kevin Swersky, and Richard Zemel. 2017. Prototypical networks for few-shot learning. In NeurIPS. 4077--4087.
[12]
Gjorgji Strezoski, Nanne van Noord, and Marcel Worring. 2019. Many task learning with task routing. In ICCV. 1375--1384.
[13]
Eric Tzeng, Judy Hoffman, Kate Saenko, and Trevor Darrell. 2017. Adversarial discriminative domain adaptation. In CVPR.
[14]
Oriol Vinyals, Charles Blundell, Timothy Lillicrap, Daan Wierstra, et almbox. 2016. Matching networks for one shot learning. In NeurIPS. 3630--3638.
[15]
Huaxiu Yao, Yiding Liu, Ying Wei, Xianfeng Tang, and Zhenhui Li. 2019 a. Learning from Multiple Cities: A Meta-Learning Approach for Spatial-Temporal Prediction. In WWW. 2181--2191.
[16]
Huaxiu Yao, Ying Wei, Junzhou Huang, and Zhenhui Li. 2019 b. Hierarchically Structured Meta-learning. In ICML. 7045--7054.
[17]
Huaxiu Yao, Xian Wu, Zhiqiang Tao, Yaliang Li, Bolin Ding, Ruirui Li, and Zhenhui Li. 2020. Automated Relational Meta-learning. In ICLR.
[18]
Jason Yosinski, Jeff Clune, Yoshua Bengio, and Hod Lipson. 2014. How transferable are features in deep neural networks?. In NeurIPS. 3320--3328.
[19]
Xi Sheryl Zhang, Fengyi Tang, Hiroko H Dodge, Jiayu Zhou, and Fei Wang. 2019. Metapred: Meta-learning for clinical risk prediction with limited patient electronic health records. In KDD.

Cited By

View all
  • (2023)Lie Group Equivariant Convolutional Neural Network Based on Laplace DistributionRemote Sensing10.3390/rs1515375815:15(3758)Online publication date: 28-Jul-2023
  • (2021)How to Handle Health-Related Small Imbalanced Data in Machine Learning?i-com10.1515/icom-2020-001819:3(215-226)Online publication date: 15-Jan-2021
  • (2021)Deep Fake Recognition in Tweets Using Text Augmentation, Word Embeddings and Deep LearningComputational Science and Its Applications – ICCSA 202110.1007/978-3-030-86979-3_37(523-538)Online publication date: 13-Sep-2021

Recommendations

Comments

Please enable JavaScript to view thecomments powered by Disqus.

Information & Contributors

Information

Published In

cover image ACM Conferences
KDD '20: Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining
August 2020
3664 pages
ISBN:9781450379984
DOI:10.1145/3394486
Permission to make digital or hard copies of part or all of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for third-party components of this work must be honored. For all other uses, contact the Owner/Author.

Sponsors

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 20 August 2020

Check for updates

Author Tags

  1. domain knowledge
  2. knowledge transfer
  3. meta-learning

Qualifiers

  • Tutorial

Funding Sources

Conference

KDD '20
Sponsor:

Acceptance Rates

Overall Acceptance Rate 1,133 of 8,635 submissions, 13%

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)33
  • Downloads (Last 6 weeks)5
Reflects downloads up to 20 Nov 2024

Other Metrics

Citations

Cited By

View all
  • (2023)Lie Group Equivariant Convolutional Neural Network Based on Laplace DistributionRemote Sensing10.3390/rs1515375815:15(3758)Online publication date: 28-Jul-2023
  • (2021)How to Handle Health-Related Small Imbalanced Data in Machine Learning?i-com10.1515/icom-2020-001819:3(215-226)Online publication date: 15-Jan-2021
  • (2021)Deep Fake Recognition in Tweets Using Text Augmentation, Word Embeddings and Deep LearningComputational Science and Its Applications – ICCSA 202110.1007/978-3-030-86979-3_37(523-538)Online publication date: 13-Sep-2021

View Options

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media