Nothing Special   »   [go: up one dir, main page]

skip to main content
10.5555/3306127.3332052acmconferencesArticle/Chapter ViewAbstractPublication PagesaamasConference Proceedingsconference-collections
research-article

The StarCraft Multi-Agent Challenge

Published: 08 May 2019 Publication History

Abstract

In the last few years, deep multi-agent reinforcement learning (RL) has become a highly active area of research. A particularly challenging class of problems in this area is partially observable, cooperative, multi-agent learning, in which teams of agents must learn to coordinate their behaviour while conditioning only on their private observations. This is an attractive research area since such problems are relevant to a large number of real-world systems and are also more amenable to evaluation than general-sum problems. Standardised environments such as the ALE and MuJoCo have allowed single-agent RL to move beyond toy domains, such as grid worlds. However, there is no comparable benchmark for cooperative multi-agent RL. As a result, most papers in this field use one-off toy problems, making it difficult to measure real progress. In this paper, we propose the StarCraft Multi-Agent Challenge (SMAC) as a benchmark problem to fill this gap. SMAC is based on the popular real-time strategy game StarCraft II and focuses on micromanagement challenges where each unit is controlled by an independent agent that must act based on local observations. We offer a diverse set of challenge maps and recommendations for best practices in benchmarking and evaluations. We also open-source a deep multi-agent RL learning framework including state-of-the-art algorithms. We believe that SMAC can provide a standard benchmark environment for years to come. Videos of our best agents for several SMAC scenarios are available at: https://youtu.be/VZ7zmQ_obZ0.

References

[1]
M. G. Bellemare, Y. Naddaf, J. Veness, and M. Bowling. 2013. The Arcade Learning Environment: An Evaluation Platform for General Agents. Journal of Artificial Intelligence Research, Vol. 47 (jun 2013), 253--279.
[2]
DeepMind. 2019. AlphaStar: Mastering the Real-Time Strategy Game StarCraft II. (2019). https://deepmind.com/blog/alphastar-mastering-real-time-strategy-game-starcraft-ii/
[3]
Jakob Foerster, Gregory Farquhar, Triantafyllos Afouras, Nantas Nardelli, and Shimon Whiteson. 2018. Counterfactual Multi-Agent Policy Gradients. In Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence .
[4]
Jakob Foerster, Nantas Nardelli, Gregory Farquhar, Triantafyllos Afouras, Philip H. S. Torr, Pushmeet Kohli, and Shimon Whiteson. 2017. Stabilising Experience Replay for Deep Multi-Agent Reinforcement Learning. In Proceedings of the 34th International Conference on Machine Learning. 1146--1155.
[5]
Johannes Heinrich and David Silver. 2016. Deep Reinforcement Learning from Self-Play in Imperfect-Information Games. CoRR, Vol. abs/1603.01121 (2016). arxiv: 1603.01121 http://arxiv.org/abs/1603.01121
[6]
Landon Kraemer and Bikramjit Banerjee. 2016. Multi-agent reinforcement learning as a rehearsal for decentralized planning. Neurocomputing, Vol. 190 (2016), 82--94.
[7]
Joel Z. Leibo, Vinicius Zambaldi, Marc Lanctot, Janusz Marecki, and Thore Graepel. 2017. Multi-agent Reinforcement Learning in Sequential Social Dilemmas. (Feb. 2017). http://arxiv.org/abs/1702.03037 arXiv: 1702.03037.
[8]
Ryan Lowe, Yi Wu, Aviv Tamar, Jean Harb, Pieter Abbeel, and Igor Mordatch. 2017. Multi-Agent Actor-Critic for Mixed Cooperative-Competitive Environments. (June 2017). http://arxiv.org/abs/1706.02275 arXiv: 1706.02275.
[9]
Frans A. Oliehoek, Matthijs T. J. Spaan, and Nikos Vlassis. 2008. Optimal and Approximate Q-value Functions for Decentralized POMDPs. JAIR, Vol. 32 (2008), 289--353.
[10]
Matthias Plappert, Marcin Andrychowicz, Alex Ray, Bob McGrew, Bowen Baker, Glenn Powell, Jonas Schneider, Josh Tobin, Maciek Chociej, Peter Welinder, Vikash Kumar, and Wojciech Zaremba. 2018. Multi-Goal Reinforcement Learning: Challenging Robotics Environments and Request for Research. (2018).

Cited By

View all
  • (2024)Explaining Sequences of Actions in Multi-agent Deep Reinforcement Learning ModelsProceedings of the 23rd International Conference on Autonomous Agents and Multiagent Systems10.5555/3635637.3663219(2537-2539)Online publication date: 6-May-2024
  • (2024)When is Mean-Field Reinforcement Learning Tractable and Relevant?Proceedings of the 23rd International Conference on Autonomous Agents and Multiagent Systems10.5555/3635637.3663068(2038-2046)Online publication date: 6-May-2024
  • (2024)Aligning Credit for Multi-Agent Cooperation via Model-based Counterfactual ImaginationProceedings of the 23rd International Conference on Autonomous Agents and Multiagent Systems10.5555/3635637.3662876(281-289)Online publication date: 6-May-2024
  • Show More Cited By

Recommendations

Comments

Please enable JavaScript to view thecomments powered by Disqus.

Information & Contributors

Information

Published In

cover image ACM Conferences
AAMAS '19: Proceedings of the 18th International Conference on Autonomous Agents and MultiAgent Systems
May 2019
2518 pages
ISBN:9781450363099

Sponsors

Publisher

International Foundation for Autonomous Agents and Multiagent Systems

Richland, SC

Publication History

Published: 08 May 2019

Check for updates

Author Tags

  1. multi-agent learning
  2. reinforcement learning
  3. starcraft

Qualifiers

  • Research-article

Conference

AAMAS '19
Sponsor:

Acceptance Rates

AAMAS '19 Paper Acceptance Rate 193 of 793 submissions, 24%;
Overall Acceptance Rate 1,155 of 5,036 submissions, 23%

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)27
  • Downloads (Last 6 weeks)3
Reflects downloads up to 16 Dec 2024

Other Metrics

Citations

Cited By

View all
  • (2024)Explaining Sequences of Actions in Multi-agent Deep Reinforcement Learning ModelsProceedings of the 23rd International Conference on Autonomous Agents and Multiagent Systems10.5555/3635637.3663219(2537-2539)Online publication date: 6-May-2024
  • (2024)When is Mean-Field Reinforcement Learning Tractable and Relevant?Proceedings of the 23rd International Conference on Autonomous Agents and Multiagent Systems10.5555/3635637.3663068(2038-2046)Online publication date: 6-May-2024
  • (2024)Aligning Credit for Multi-Agent Cooperation via Model-based Counterfactual ImaginationProceedings of the 23rd International Conference on Autonomous Agents and Multiagent Systems10.5555/3635637.3662876(281-289)Online publication date: 6-May-2024
  • (2024)Enhancing Multi-agent System Testing with Diversity-Guided Exploration and Adaptive Critical State ExploitationProceedings of the 33rd ACM SIGSOFT International Symposium on Software Testing and Analysis10.1145/3650212.3680376(1491-1503)Online publication date: 11-Sep-2024
  • (2024)Fuzzy Feedback Multiagent Reinforcement Learning for Adversarial Dynamic Multiteam CompetitionsIEEE Transactions on Fuzzy Systems10.1109/TFUZZ.2024.336305332:5(2811-2824)Online publication date: 7-Feb-2024
  • (2023)Offline multi-agent reinforcement learning with implicit global-to-local value regularizationProceedings of the 37th International Conference on Neural Information Processing Systems10.5555/3666122.3668404(52413-52429)Online publication date: 10-Dec-2023
  • (2023)SMACv2Proceedings of the 37th International Conference on Neural Information Processing Systems10.5555/3666122.3667756(37567-37593)Online publication date: 10-Dec-2023
  • (2023)Policy mirror ascent for efficient and independent learning in mean field gamesProceedings of the 40th International Conference on Machine Learning10.5555/3618408.3620066(39722-39754)Online publication date: 23-Jul-2023
  • (2023)Attention-based recurrence for multi-agent reinforcement learning under stochastic partial observabilityProceedings of the 40th International Conference on Machine Learning10.5555/3618408.3619565(27840-27853)Online publication date: 23-Jul-2023
  • (2022)Rethinking individual global max in cooperative multi-agent reinforcement learningProceedings of the 36th International Conference on Neural Information Processing Systems10.5555/3600270.3602620(32438-32449)Online publication date: 28-Nov-2022
  • Show More Cited By

View Options

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media