Publication: Encouraging Cooperation in Multi Agent Reinforcement Learning
No Thumbnail Available
Open/View Files
Date
2023-06-30
Authors
Published Version
Published Version
Journal Title
Journal ISSN
Volume Title
Publisher
The Harvard community has made this article openly available. Please share how this access benefits you.
Citation
Kwun, Mujin. 2023. Encouraging Cooperation in Multi Agent Reinforcement Learning. Bachelor's thesis, Harvard College.
Research Data
Abstract
Encouraging cooperation in Multi-agent reinforcement learning (MARL) remains a big area of research. In addition, additional complexity as well as non-stationarity when scaling up from the single-agent setting makes convergence to optimal policies difficult compared to single-agent reinforcement learning. In this thesis, we build on previous work demonstrating the empirical effectiveness of policy-gradient methods in multi-agent settings, specifically Proximal Policy Optimization(PPO). We introduce a novel test-bed for multi-agent reinforcement learning and evaluate the effectiveness of a decentralized PPO framework in this test-bed. Furthermore, motivated by literature that shows the benefits of reward shaping on convergence in the single agent setting, we apply domain specific reward shaping to our PPO method to encourage cooperation and faster convergence to a good joint policy.
Description
Other Available Sources
Keywords
Statistics
Terms of Use
This article is made available under the terms and conditions applicable to Other Posted Material (LAA), as set forth at Terms of Service