log in  |  register  |  feedback?  |  help  |  web accessibility
Logo
MAESTRO: Open-Ended Environment Design for Multi-Agent Reinforcement Learning
Tuesday, May 9, 2023, 4:00-5:00 pm Calendar
  • You are subscribed to this talk through .
  • You are watching this talk through .
  • You are subscribed to this talk. (unsubscribe, watch)
  • You are watching this talk. (unwatch, subscribe)
  • You are not subscribed to this talk. (watch, subscribe)
Registration requested: The organizer of this talk requests that you register if you are planning to attend. There are two ways to register: (1) You can create an account on this site (click the "register" link in the upper-right corner) and then register for this talk; or (2) You can enter your details below and click the "Register for talk" button. Either way, you can always cancel your registration later.

Name:
Email:
Organization:

Abstract

Open-ended learning methods that automatically generate a curriculum of increasingly challenging tasks serve as a promising avenue toward generally capable reinforcement learning agents. Existing methods adapt curricula independently over either environment parameters (in single-agent settings) or co-player policies (in multi-agent settings). However, the strengths and weaknesses of co-players can manifest themselves differently depending on environmental features. It is thus crucial to consider the dependency between the environment and co-player when shaping a curriculum in multi-agent domains. In this work, we use this insight and extend Unsupervised Environment Design (UED) to multi-agent environments. We then introduce Multi-Agent Environment Design Strategist for Open-Ended Learning (MAESTRO), the first multi-agent UED approach for two-player zero-sum settings. MAESTRO efficiently produces adversarial, joint curricula over both environments and co-players and attains minimax-regret guarantees at Nash equilibrium. Our experiments show that MAESTRO outperforms a number of strong baselines on competitive two-player games, spanning discrete and continuous control settings.

 

Bio

Mikayel Samvelyan is a Research Assistant at Meta AI and a PhD student at University College London. At UCL, he is supervised by Tim Rocktäschel and is a member of the UCL DARK Lab, as well as part of the ELLIS PhD & Postdoc Program. Mikayel holds an MSc in Computer Science degree from the University of Oxford, where he was advised by Shimon Whiteson. His research focuses on deep reinforcement learning, multi-agent systems, and open-ended learning.

This talk is organized by Saptarashmi Bandyopadhyay