Keywords: population learning, emergent communication
TL;DR: We propose an adaptive population-based method to train agents in a cooperative multi-agent reinforcement learning setup. We show that our method induces useful diversity into a population of agents that helps in learning a more robust meta-agent.
Abstract: Recent works have shown remarkable progress in training artificial agents to understand natural language but are focused on using large amounts of raw data involving huge compute requirements. An interesting hypothesis follows the idea of training artificial agents via multi-agent communication while using small amounts of task-specific human data to ground the emergent language into natural language. This allows agents to communicate with humans without needing enormous expensive human demonstrations. Evolutionary studies have showed that simpler and easily adaptable languages arise as a result of communicating with a diverse group of large population. We propose to model this supposition with artificial agents and propose an adaptive population-based meta-reinforcement learning approach that builds such a population in an iterative manner. We show empirical results on referential games involving natural language where our agents outperform all baselines on both the task performance and language score including human evaluation. We demonstrate that our method induces constructive diversity into a growing population of agents that is beneficial in training the meta-agent.
Proposed Reviewers: Abhinav Gupta, abhinav.gupta@umontreal.ca
0 Replies
Loading