Chen-yu Wei (University of Southern California)
Room 116 and Zoom
Reinforcement learning is typically built upon the assumption that the environment is fixed and uncorrupted. This assumption does not hold anymore when there exists adversarial corruption, non-stationary transition, or other agents’ time-varying policies. Some standard reinforcement learning algorithms are vulnerable to these factors – e.g., a tiny amount of corruption may totally alter the behavior of the algorithm. In the first part of the talk, we will discuss reduction techniques that turn a standard algorithm that only works for stationary environments into one that is robust to non-stationarity. The reductions are black-box, general, and optimal for a wide range of problems.
In the second part, we will focus on decentralized multi-agent reinforcement learning, in which each agent also faces a non-stationary environment. Decentralized algorithms are easy to implement, versatile for different types of games, and scalable to systems with many agents, but they often suffer from non-convergence issues. We will discuss algorithmic techniques that facilitate the convergence of the system, while not introducing extra coordination or communication overhead.
Bio: Chen-Yu Wei is a Computer Science PhD candidate at University of Southern California, advised by Haipeng Luo. His research focuses on online decision making, reinforcement learning, and learning in games. He is a recipient of the Best Paper Award at Conference on Learning Theory (COLT) 2021, and the Best Research Assistant Award at the Department of Computer Science, University of Southern California in 2020.