Nothing Special   »   [go: up one dir, main page]

skip to main content
10.5555/2074284.2074297guideproceedingsArticle/Chapter ViewAbstractPublication PagesConference Proceedingsacm-pubtype
Article
Free access

Learning conventions in multiagent stochastic domains using likelihood estimates

Published: 01 August 1996 Publication History

Abstract

Fully cooperative multiagent systems--those in which agents share a joint utility model--is of special interest in AI. A key problem is that of ensuring that the actions of individual agents are coordinated, especially in settings where the agents are autonomous decision makers. We investigate approaches to learning coordinated strategies in stochastic domains where an agent's actions are not directly observable by others. Much recent work in game theory has adopted a Bayesian learning perspective to the more general problem of equilibrium selection, but tends to assume that actions can be observed. We discuss the special problems that arise when actions are not observable, including effects on rates of convergence, and the effect of action failure probabilities and asymmetries. We also use likelihood estimates as a means of generalizing fictitious play learning models in our setting. Finally, we propose the use of maximum likelihood as a means of removing strategies from consideration, with the aim of convergence to a conventional equilibrium, at which point learning and deliberation can cease.

References

[1]
Robert Axelrod. The Evolution of Cooperation. Basic Books, New York, 1984.
[2]
Craig Boutilier. Planning, learning and coordination in multiagent decision processes. In Proceedings of the Sixth Conference on Theoretical Aspects of Rationality and Knowledge, Amsterdam, 1996. (to appear).
[3]
Craig Boutilier, Richard Dearden, and Moisés Goldszmidt. Exploiting structure in policy construction. In Proceedings of the Fourteenth International Joint Conference on Artificial Intelligence, pages 1104-1111, Montreal, 1995.
[4]
Drew Fudenberg and David K. Levine, Steady state learning and hash.equilibrium. Econometrica, 61(3):547-573, 1993.
[5]
John C. Harsanyi and Reinhard Selten. A General Theory of Equilibrium Selection in Games. MIT Press, Cambridge, 1988.
[6]
David Heckerman. A tutorial on learning bayesian networks. Technical Report MSR-TR-95-06, Microsoft Research, Redmond, WA, 1995.
[7]
Ronald A. Howard. Dynamic Programming and Markov Processes. MIT Press, Cambridge, 1960.
[8]
Ehud Kalai and Ehud Lehrer. Rational learning leads to nash equilibrium. Econometrica, 61(5):1019-1045, 1993.
[9]
David K. Lewis. Conventions, A Philosophical Study. Harvard University Press, Cambridge, 1969.
[10]
Michael L. Littman. Markov games as a framework for multi-agent reinforcement learning. In Proceedings of the Eleventh International Conference on Machine Learning, pages 157-163, New Brunswick, NJ, 1994.
[11]
George Mailath, Michihiro Kandori and Rafael Rob. Learning, mutation and long run equilibria in games. Econometrica, 61(1):29-56, 1993.
[12]
Roger B. Myerson. Game Theory: Analysis of Conflict. Harvard University Press, Cambridge, 1991.
[13]
Guillermo Owen. Game Theory. Academic Press, New York, 1982.
[14]
Martin L. Puterman. Markov Decision Processes: Discrete Stochastic Dynamic Programming. Wiley, New York, 1994.
[15]
Yoav Shoham and Moshe Tennenholtz. Emergent conventions in multi-agent systems: Initial experimental results and observations. In Third Intl. Conf. on Principles of Knowledge Representation and Reasoning, pages 225-231, Cambridge, 1992.
[16]
Yoav Shoham and Moshe Tenoenholtz. On the synthesis of useful social laws for artificial agent societies. In Proceedings of the Tenth National Conference on Artificial Intelligence, pages 276-281, San Jose, 1992.
[17]
Gerhard Weiß Learning to coordinate actions in multi-agent systems. In Proceedings of the Thirteenth International Joint Conference on Artificial Intelligence, pages 311-316, Chambery, FR, 1993.
[18]
Holly Yanco and Lynn Andrea Stein. An adaptive communication protocol for cooperating mobile robots. Proc. of the Second International Conference on the Simulation of Adaptive Behavior, pages 478-485. MIT Press, Cambridge, 1993.
[19]
H. Peyton Young. The evolution of conventions. Econometrica, 61(1):57-84, 1993.

Cited By

View all
  • (2017)Multi-agent actor-critic for mixed cooperative-competitive environmentsProceedings of the 31st International Conference on Neural Information Processing Systems10.5555/3295222.3295385(6382-6393)Online publication date: 4-Dec-2017
  • (2017)Learning Conventions via Social Reinforcement Learning in Complex and Open SettingsProceedings of the 16th Conference on Autonomous Agents and MultiAgent Systems10.5555/3091125.3091193(455-463)Online publication date: 8-May-2017
  • (2001)A Model of Partially Observable State Game and its OptimalityApplied Intelligence10.1023/A:101129471985214:3(273-284)Online publication date: 9-May-2001
  • Show More Cited By

Recommendations

Comments

Please enable JavaScript to view thecomments powered by Disqus.

Information & Contributors

Information

Published In

cover image Guide Proceedings
UAI'96: Proceedings of the Twelfth international conference on Uncertainty in artificial intelligence
August 1996
572 pages
ISBN:155860412X

Sponsors

  • Ricoh California Research Center: Ricoh California Research Center
  • Rockwell Science Center: Rockwell Science Center
  • HUGIN: Hugin Expert A/S
  • Information Extraction and Transportation
  • Microsoft Research: Microsoft Research

Publisher

Morgan Kaufmann Publishers Inc.

San Francisco, CA, United States

Publication History

Published: 01 August 1996

Qualifiers

  • Article

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)36
  • Downloads (Last 6 weeks)11
Reflects downloads up to 22 Nov 2024

Other Metrics

Citations

Cited By

View all
  • (2017)Multi-agent actor-critic for mixed cooperative-competitive environmentsProceedings of the 31st International Conference on Neural Information Processing Systems10.5555/3295222.3295385(6382-6393)Online publication date: 4-Dec-2017
  • (2017)Learning Conventions via Social Reinforcement Learning in Complex and Open SettingsProceedings of the 16th Conference on Autonomous Agents and MultiAgent Systems10.5555/3091125.3091193(455-463)Online publication date: 8-May-2017
  • (2001)A Model of Partially Observable State Game and its OptimalityApplied Intelligence10.1023/A:101129471985214:3(273-284)Online publication date: 9-May-2001
  • (1999)Sequential optimality and coordination in multiagent systemsProceedings of the 16th international joint conference on Artifical intelligence - Volume 110.5555/1624218.1624287(478-485)Online publication date: 31-Jul-1999
  • (1998)Conjectural Equilibrium in Multiagent LearningMachine Language10.1023/A:100751462358933:2-3(179-200)Online publication date: 1-Dec-1998

View Options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Login options

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media