A Study of AI Population Dynamics with Million-agent Reinforcement Learning

Published in The 17th International Conference on Autonomous Agents and Multiagent Systems (Extended Abstract) & NIPS-17 Aligned Artificial Intelligence Workshop, 2017

Lantao Yu*, Yaodong Yang*, Yiwei Bai*(equal contribution), Jun Wang, Weinan Zhang, Ying Wen, Yong Yu. The 17th International Conference on Autonomous Agents and Multiagent Systems (Extended Abstract) & NIPS-17 Aligned Artificial Intelligence Workshop.

[PDF] [Code]

Abstract

In this paper, we conduct an empirical study on discovering the ordered collective dynamics obtained by a population of artificial intelligence (AI) agents. Our intention is to put AI agents into a simulated natural context, and then to understand their induced dynamics at the population level. In particular, we aim to verify if the principles developed in the real world could also be used in understanding an artificially-created intelligent population. To achieve this, we simulate a large-scale predator-prey world, where the laws of the world are designed by only the findings or logical equivalence that have been discovered in nature. We endow the agents with the intelligence based on deep reinforcement learning, and scale the population size up to millions. Our results show that the population dynamics of AI agents, driven only by each agent’s individual self interest, reveals an ordered pattern that is similar to the Lotka-Volterra model studied in population biology. We further discover the emergent behaviors of collective adaptations in studying how the agents’ grouping behaviors will change with the environmental resources. Both of the two findings could be explained by the self-organization theory in nature.