Timezone: »
Inspired by organisms evolving through cooperation and competition between different populations on Earth, we study the emergence of artificial collective intelligence through massive-agent reinforcement learning. To this end, We propose a new massive-agent reinforcement learning environment, Lux, where dynamic and massive agents in two teams scramble for limited resources and fight off the darkness. In Lux, we build our agents through the standard reinforcement learning algorithm in curriculum learning phases and leverage centralized control via a pixel-to-pixel policy network. As agents co-evolve through self-play, we observe several stages of intelligence, from the acquisition of atomic skills to the development of group strategies. Since these learned group strategies arise from individual decisions without an explicit coordination mechanism, we claim that artificial collective intelligence emerges from massive-agent cooperation and competition. We further analyze the emergence of various learned strategies through metrics and ablation studies, aiming to provide insights for reinforcement learning implementations in massive-agent environments.
Author Information
Hanmo Chen
Stone Tao (University of California - San Diego)
JIAXIN CHEN (The Hong Kong Polytechnic University)
Weihan Shen (University of Florida)
Xihui Li (Tsinghua University, Tsinghua University)
Chenghui Yu
Sikai Cheng (Georgia Institute of Technology)
Xiaolong Zhu (University of Hong Kong)
Xiu Li
More from the Same Authors
-
2021 Spotlight: Overcoming Catastrophic Forgetting in Incremental Few-Shot Learning by Finding Flat Minima »
Guangyuan SHI · JIAXIN CHEN · Wenlong Zhang · Li-Ming Zhan · Xiao-Ming Wu -
2021 : ManiSkill: Generalizable Manipulation Skill Benchmark with Large-Scale Demonstrations »
Tongzhou Mu · Zhan Ling · Fanbo Xiang · Derek Yang · Xuanlin Li · Stone Tao · Zhiao Huang · Zhiwei Jia · Hao Su -
2021 : MHER: Model-based Hindsight Experience Replay »
Yang Rui · Meng Fang · Lei Han · Yali Du · Feng Luo · Xiu Li -
2022 Poster: Double Check Your State Before Trusting It: Confidence-Aware Bidirectional Offline Model-Based Imagination »
Jiafei Lyu · Xiu Li · Zongqing Lu -
2022 Poster: OrdinalCLIP: Learning Rank Prompts for Language-Guided Ordinal Regression »
Wanhua Li · Xiaoke Huang · Zheng Zhu · Yansong Tang · Xiu Li · Jie Zhou · Jiwen Lu -
2022 Poster: Mildly Conservative Q-Learning for Offline Reinforcement Learning »
Jiafei Lyu · Xiaoteng Ma · Xiu Li · Zongqing Lu -
2022 : State Advantage Weighting for Offline RL »
Jiafei Lyu · aicheng Gong · Le Wan · Zongqing Lu · Xiu Li -
2022 : Fifteen-minute Competition Overview Video »
Joseph Suarez · Hanmo Chen -
2022 : Abstract-to-Executable Trajectory Translation for One-Shot Task Generalization »
Stone Tao · Xiaochen Li · Tongzhou Mu · Zhiao Huang · Yuzhe Qin · Hao Su -
2022 Competition: The Third Neural MMO Challenge: Learning to Specialize in Massively Multiagent Open Worlds »
Joseph Suarez · Hanmo Chen · Arbin Chen · Bo Wu · Xiaolong Zhu · enhong liu · JUN HU · Chenghui Yu · Phillip Isola -
2022 Spotlight: Mildly Conservative Q-Learning for Offline Reinforcement Learning »
Jiafei Lyu · Xiaoteng Ma · Xiu Li · Zongqing Lu -
2022 Spotlight: Double Check Your State Before Trusting It: Confidence-Aware Bidirectional Offline Model-Based Imagination »
Jiafei Lyu · Xiu Li · Zongqing Lu -
2021 Poster: Overcoming Catastrophic Forgetting in Incremental Few-Shot Learning by Finding Flat Minima »
Guangyuan SHI · JIAXIN CHEN · Wenlong Zhang · Li-Ming Zhan · Xiao-Ming Wu -
2020 Poster: A Closer Look at the Training Strategy for Modern Meta-Learning »
JIAXIN CHEN · Xiao-Ming Wu · Yanke Li · Qimai LI · Li-Ming Zhan · Fu-lai Chung