Nothing Special   »   [go: up one dir, main page]

skip to main content
10.5555/3523760.3523964acmconferencesArticle/Chapter ViewAbstractPublication PageshriConference Proceedingsconference-collections
research-article

Personalized Meta-Learning for Domain Agnostic Learning from Demonstration

Published: 07 March 2022 Publication History

Abstract

For robots to perform novel tasks in the real-world, they must be capable of learning from heterogeneous, non-expert human teachers across various domains. Yet, novice human teachers often provide suboptimal demonstrations, making it difficult for robots to successfully learn. Therefore, to effectively learn from humans, we must develop learning methods that can account for teacher suboptimality and can do so across various robotic platforms. To this end, we introduce Mutual Information Driven Meta-Learning from Demonstration (MIND MELD) [12, 13], a personalized meta-learning framework which meta-learns a mapping from suboptimal human feedback to feedback closer to optimal, conditioned on a learned personalized embedding. In a human subjects study, we demonstrate MIND MELD's ability to improve upon suboptimal demonstrations and learn meaningful, personalized embeddings. We then propose Domain Agnostic MIND MELD, which learns to transfer the personalized embedding learned in one domain to a novel domain, thereby allowing robots to learn from suboptimal humans across disparate platforms (e.g., self-driving car or in-home robot).

References

[1]
Saleema Amershi, Maya Cakmak, W. Bradley Knox, and Todd Kulesza. Power to the people: The role of humans in interactive machine learning. AI Magazine, 35(4):105--120, 2014.
[2]
Brenna Argall, Sonia Chernova, Manuela M. Veloso, and Brett Browning. A survey of robot learning from demonstration. Robotics and Autonomous Systems, 57(5):469--483, May 2009.
[3]
Sonia Chernova and Manuela Veloso. Interactive policy learning through confidence-based autonomy. Journal of Artificial Intelligence Research, 34:1--25, 2009.
[4]
Muhammad Abdullah Jamal and Guo-Jun Qi. Task agnostic meta-learning for few-shot learning.
[5]
Michael Laskey, Caleb Chuck, Jonathan Lee, Jeffrey Mahler, Sanjay Krishnan, Kevin Jamieson, Anca Dragan, and Ken Goldberg. Comparing human-centric and robot-centric sampling for robot deep learning from demonstrations. Proceedings - IEEE International Conference on Robotics and Automation, pages 358--365, 2017.
[6]
Takayuki Osa, Gerhard Neumann, and Jan Peters. An Algorithmic Perspective on Imitation Learning. 7(1):1--179, 2018.
[7]
Rohan Paleja and Matthew Gombolay. Inferring personalized bayesian embeddings for learning from heterogeneous demonstration. arXiv, 2019.
[8]
Laurel D. Riek. Wizard of oz studies in hri: A systematic review and new reporting guidelines. J. Hum.-Robot Interact., 1(1):119--136, July 2012.
[9]
Sté phane Ross, Geoffrey J Gordon, and J. Andrew Bagnell. No-regret reductions for imitation learning and structured prediction. Aistats, 15:627--635, 2011.
[10]
Claude Sammut. Automatically Constructing Control Systems by Observing Human Behaviour. Second International Inductive Logic Programming Workshop, (May), 1992.
[11]
Stefan Schaal. Learning from demonstration. In M. C. Mozer, M. Jordan, and T. Petsche, editors, Advances in Neural Information Processing Systems, volume 9. MIT Press, 1997.
[12]
Mariah L. Schrum, Nina Moorman Erin Hedlund, and Matthew C. Gombolay. Mind meld: Personalized meta-learning for robot-centric imitation learning. ACM/IEEE International Conference on Human-Robot Interaction, 2022.
[13]
Mariah L. Schrum, Erin Hedlund, and Matthew C. Gombolay. Improving robot-centric learning from demonstration via personalized embeddings. CoRR, abs/2110.03134, 2021.
[14]
Felipe Leno Da Silva, Garrett Warnell, Anna Helena Reali Costa, and Peter Stone. Agents teaching agents: a survey on inter-agent transfer learning. Autonomous Agents and Multi-Agent Systems, 34, 4 2020.
[15]
Jonathan Spencer, Sanjiban Choudhury, Matt Barnes, Matthew Schmittle, Mung Chiang, Peter Ramadge, and Siddhartha Srinivasa. Learning from Interventions: Human-robot interaction as both explicit and implicit feedback. 2020.
[16]
Sebastian Weigelt, Vanessa Steurer, and Walter F. Tichy. At your command! an empirical study on how laypersons teach robots new functions. In 2020 IEEE 14th International Conference on Semantic Computing (ICSC), pages 468--470, 2020.

Recommendations

Comments

Please enable JavaScript to view thecomments powered by Disqus.

Information & Contributors

Information

Published In

cover image ACM Conferences
HRI '22: Proceedings of the 2022 ACM/IEEE International Conference on Human-Robot Interaction
March 2022
1353 pages

Sponsors

Publisher

IEEE Press

Publication History

Published: 07 March 2022

Check for updates

Author Tags

  1. meta-learning
  2. personalized learning

Qualifiers

  • Research-article

Funding Sources

  • National Science Foundation
  • Konica Minolta Inc
  • NASA Early Career Fellowship
  • MIT Lincoln Laboratory

Conference

HRI '22
Sponsor:

Acceptance Rates

Overall Acceptance Rate 268 of 1,124 submissions, 24%

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • 0
    Total Citations
  • 123
    Total Downloads
  • Downloads (Last 12 months)17
  • Downloads (Last 6 weeks)3
Reflects downloads up to 30 Nov 2024

Other Metrics

Citations

View Options

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media